Automatically Distinguishing People’s Explicit and Implicit Attitude Bias by Bridging Psychological Measurements with Sentiment Analysis on Large Corpora

: Social psychological studies show that people’s explicit attitude bias in public expression can differ from their implicit attitude bias in mind. However, the current automatic attitude analysis does not distinguish between explicit and implicit attitude bias. Simulating the psychological measurements of explicit and implicit attitude bias, i.e., self-report assessment (SRA) and implicit association test (IAT), we propose an automatic language-based analysis to distinguish explicit and implicit attitude bias in a large population. By connecting the criteria of SRA and IAT with the statements containing patterns of special words, we derive explicit and implicit attitude bias with the sentiment scores of the statements, which are obtained by pre-trained machine-learning methods. Extensive experiments on four English and Chinese corpora and four pairs of concepts show that the attitude biases obtained by our method on a large population are consistent with those of traditional psychological experiments in the costly small-scale experiments. The maximum gap between the sentiment scores of explicit and implicit biases reaches 0.9329. Furthermore, we achieve new findings on the difference between the evolution of explicit and implicit attitude bias. The maximum variance gap of sentiment scores in the dynamic changes between explicit and implicit biases reaches 0.249.


Introduction
Attitude bias (abbreviated as "bias") is people's tendentious beliefs about concepts.As an inherent motivation of people's interests and consequent behaviors, bias is essential to human-centered research, such as social psychology, computational opinion analysis, and recommendation.In current natural-language processing research, we can understand biases by analyzing the sentiment and semantics of people's public expressions.However, social psychology research reveals that people's explicitly expressed bias in public does not always agree with their implicit bias in mind [1].People's explicit biases are conscious, controllable, and easy to report, while implicit biases are uncontrollable and cannot be consciously acquired [2].Explicit and implicit biases play different roles in social life.Explicit bias spreads in public and forms the mainstream value, whereas implicit bias can determine the behavior without conscious awareness, e.g., voting and purchasing [3].A well-known example is American racial bias, where self-reports reveal a near-absence of preference difference between White and Black people [4].In contrast, implicit bias reveals a widespread preference for White people relative to Black people [5].Implicit bias is also found to be a better predictor of certain kinds of behavior (e.g., risky flight [6], and anxiety [7]) than explicit bias.
In psychology, the Self-Report Assessment (SRA) [8] and the Implicit Association Test (IAT) [5] are capable of measuring explicit and implicit attitude bias, respectively.However, both SRA and IAT need cooperation from subjects, and it is very costly for a large population.To understand the bias automatically, researchers apply natural-language processing techniques to understand bias with the semantic relations between people's words [9].For example, Recasents [10] introduced that biased language in Wikipedia uncovers framing and epistemological bias and can be identified by familiar linguistic cues.Recent works show that word and sentence embedding can capture common bias in training corpora [11].Word Embedding Association Test (WEAT) proposes to [12] measure the bias by regarding the semantic distance between pre-trained vectors of certain words as the implicit association strength in IAT.Sentence Encoder Association Test (SEAT) [13] extends the WEAT by constructing sentences containing specific words so that the word encoders (word2vec and GloVe) in WEAT can be replaced by the sentence encoders (ELMo and BERT).Recent work [14,15] explores the subjects of word embedding biases and finds that these biases are centered around males.Moreover, Dobrzeniecka et al. propose hierarchical Bayesian modeling (HBM) [16] as an alternative to WEAT.HBM scrutinizes biases in word embeddings across various levels of granularity and yields robust results in bias assessments concerning gender, race, and religion.However, WEAT, SEAT, and earlier language-based approaches do not distinguish explicit and implicit biases.Without this distinguishment, if explicit and implicit biases contained in language are different, the bias identified by current methods will be a confusion of explicit and implicit biases.
Against the background mentioned above, the primary research motivation of this paper is to draw inspiration from the psychological measurement of biases and employ computational techniques to address the problem of separately identifying explicit and implicit bias in a large population automatically.
Based on this motivation, our research objective is to propose a novel strategy to connect the key factors in SRA and IAT with special words and statements and automatically simulate SRA and IAT with sentiment analysis of the statements.The framework is shown in Figure 1.First, we define the categories of special words named concept words, exemplar words and attribute words, corresponding to the concept, exemplar and attribute in SRA and IAT.Second, we simulate explicit response in SRA and implicit response in IAT with statements containing different combinations of special words, respectively.Finally, by measuring the sentiment of the statements, we derive explicit and implicit bias separately.In experiments on corpora of different languages (English and Chinese) and different types (Wikipedia and Social Media), our method successfully reproduces the well-known observations of SRA and IAT: (1) people's explicit and implicit biases can be different, especially on well-known cases reported in psychological studies, e.g., racial bias; (2) the difference degree is related to the sociality of the concepts, the scenario of language use, and the cultural background.Furthermore, we also find that implicit biases are more stable than explicit biases over time, and the stability is also related to social factors.
We summarize our contributions as follows: (1) Simulate psychological tests, i.e., SRA and IAT, with automatic language-based analysis, and do not acquire extra modification on the state-of-the-art natural-language processing techniques.(2) Automatically reproduce the small-scale psychological observations on a large population.
(3) Achieve new findings, indicating further advantages of the proposed automatic method compared with psychological tests.

Related Works
In this section, we will introduce the main related works in the fields of psychology and computer technology that are pertinent to the present study.These primarily include methods in psychology for measuring explicit and implicit biases, as well as word vector techniques and sentiment polarity measurement techniques in computer technology that can be used to assess biases.

Psychological Tests of Bias
We commence by introducing the SRA and IAT methods, which are utilized in psychology to measure explicit and implicit biases.
Self-report assessment (SRA) is a typical method of measuring explicit bias.Reading questions and choosing responses in the assessment is the most direct method of asking about one's feelings, prejudices, beliefs, and other information [17].SRA can be a survey, questionnaire, or poll in which subjects express their attitude explicitly.The main strength of SRA is allowing participants to describe their own experiences rather than inferring this from observing them.
SRA is more reliable for explicit bias than implicit bias [18].However, SRA also has specific disadvantages because of the way subjects behave [19].There are two main reasons for the untruthful response in SRA.The first reason is "social desirability bias", in which the subjects wish to present themselves in a socially acceptable manner.The second reason is the "lack of awareness" that the subjects are not aware of their genuine attitude in mind, which is also known as "unconscious attitude" [20].
There are also other reasons leading to untruthful responses in SRA: (1) The order of questions: response can be influenced by the different orders of questions.(2) Response options: The response of a given option can be influenced by the compared options.(3) The wording of questions: response also can be influenced by the description of the questions.
These disadvantages can be leading because they may unwittingly force the respondent to give a particular reply [21].Therefore, we will not be collecting valid data based on the explicit attitude expressed by SRA.Psychologists have developed indirect measurements, such as the Implicit Association Test (IAT), to measure implicit bias.
Implicit Association Test(IAT) [5] is the most widely used psychological test of implicit bias.In contrast to SRA, in IAT, to avoid conscious explicit bias of the concept, the test is based on the exemplars of the concept (e.g., rose) instead of the concept itself (e.g., Flower).A set of exemplar stimuli (e.g., pictures or names of exemplars) and attribute stimuli are used to prime the concept and attribute in subjects' minds unconsciously.The subjects are encouraged to classify "exemplar stimuli + attribute stimuli" combinations.During the classification, their response (e.g., button click) latency measures the implicit association between the concepts w.r.t the attributes.Finally, an effect size of response latency is used to calculate the implicit bias [22].
For example, as shown in Figure 1, in computer-aided IAT https://implicit.harvard.edu/implicit/takeatest.html (accessed on 12 March 2024), to measure the implicit bias toward "African-American vs. European-American", given two keys on the left and right hand, respectively, an "exemplar stimuli + attribute stimuli" combination is shown to the subject every time.The left screenshot pairs 'African-American' with 'Good' and 'European-American' with 'Bad,' while the right screenshot does the opposite.In the center of each screenshot is an exemplar or attribute stimuli.If the subject's response under the instruction: "click left when you see an African-American or a pleasant stimulus; click right when you see a European-American or an unpleasant stimulus" is faster than his response under the instruction: "click left when you see a European-American or a pleasant stimulus; click right when you see an African-American or an unpleasant stimulus", IAT will say that the subject has an implicit preference for "African-American" compared with "European-American".
Figure 2 illustrates a schematic diagram of a real scene where a subject is taking an IAT test in front of a computer.On the left, 'Black' is associated with 'good' and 'White' with 'bad,'.However, on the right, the associations are inverted.The test measures the subject's reaction times to these pairings to evaluate implicit racial attitudes.The following brief introduction of the Implicit Association Test (IAT) is selected from the home page of the IAT project: https://implicit.harvard.edu/implicit/index.jsp (accessed on 12 March 2024).On this website, users can try an IAT test online to learn more information.
The IAT measures the strength of associations between concepts (e.g., African-American, European-American) and attributes (e.g., good, bad) or stereotypes (e.g., athletic, clumsy).The main idea is that making a response is more accessible when closely related items share the same response key.
When doing an IAT, you are asked to quickly sort words into categories that are on the left and right-hand side of the computer screen by pressing the "e" key if the word belongs to the category on the left and the "i" key if the word belongs to the category on the right.The IAT has five main parts [23].
In the first part of the IAT, you sort words relating to the concepts (e.g., African-American, European-American) into categories.If the category "African-American" was on the left, and a picture of a Black person was on the screen, you would press the "e" key.
In the second part of the IAT, you sort words relating to the attributes (e.g., good, bad).So, if the category "good" was on the left, and a pleasant word appeared on the screen, you would press the "e" key.
In the third part of the IAT, the categories are combined, and you are asked to sort both concept and evaluation words.So, the categories on the left-hand side would be African-American/Good, and the categories on the right-hand side would be European-American/Bad.It is important to note that the order in which the blocks are presented varies across participants.Some people will do the African-American/Good, European-American/Bad part first, and others will do the African-American/Bad, European-American/Good part first.
The placement of the concepts switches in the fourth part of the IAT.If the category "African-American" were previously on the left, it would now be on the right.Notably, the number of trials in this part of the IAT is increased to minimize the effects of the practice.
In the final part of the IAT, the categories are combined in a way that is opposite to what they were before.If the category on the left were previously African-American/Good, it would now be African-American/Bad.
The IAT score is based on how long it takes a person, on average, to sort the words in the third part of the IAT versus the fifth part of the IAT.We would say that one has an implicit preference for thin people relative to fat people if they are faster to categorize words when Thin People and Good share a response key and Fat People and Bad share a response key relative to the reverse.
Although the Implicit Association Test (IAT) is currently the mainstream method for measuring implicit biases in psychological research, it has certain limitations when studying implicit biases in large-scale populations.First, the IAT requires participants to actively cooperate in the test in a laboratory or designated website, which raises the threshold for testing.Second, as participants are aware that they are being tested, the IAT falls into the category of contract testing, potentially affecting the test results.If we hope to conduct more accurate testing on a large-scale population, it is necessary to explore an automated and non-contact testing method.

Embedding Association Test of Bias
Next, we introduce a commonly used computer technology that can measure biases, embedding association test (WEAT) [12].Embedding is a semantic representation of words or sentences according to their context in the corpus [24].Words that are closer in embedding space are supposed to be semantically similar, assuming that speakers' mental associations between concepts and attributes can be represented by the semantic similarity between the words naming the concepts and attributes.Word Embedding Association Test (WEAT) [12] is proposed to measure the bias with the distance between the embeddings of the words naming the exemplars and attributes in IAT.
The details of the WEAT are as follows.It calculates the implicit association between two concept words and two sets of attribute words using Equations ( 1) and (2).Given two exemplar words set Ew i , Ew j corresponding to two concepts C i , C j , respectively (e.g., {orchid, rose, . . .} for the concept "Flower" and {ant, spider, . . .} for the concept "Insect"), and two attribute words set Aw p , Aw q corresponding to two attributes A p , A q , respectively (e.g., {happy, great, pretty, . . .} for the attribute "Pleasant" and {awful, ugly, sad, . . .} for the attribute "Unpleasant"), the null hypothesis is that there is no difference between Ew i , Ew j in terms of their relative similarity to Aw p , Aw q .Let ⃗ x be the embedding vector of word x, cos(⃗ x,⃗ y) denotes the cosine similarity between ⃗ x and ⃗ y.Specifically, WEAT [12] calculated the bias on C i , C j with respect to A p , A q as: Bias(Ew i , Ew j , Aw p , Aw q ) = ∑ e∈Ew i s(e, Aw p , Aw q ) − ∑ e∈Ew j s(e, Aw p , Aw q ) (1) where s(e, Aw p , Aw q ) =mean a∈Aw p cos(⃗ e,⃗ a) − mean a∈Aw q cos(⃗ e,⃗ a) (2) Then, WEAT [12] measures the strength of association through the effect size.The effect size ('ES' in Equation (3)) of the bias is calculated as: s(e,Aw p ,Aw q )−mean e∈Ew i s(e,Aw p ,Aw q ) std_dev e∈Ew i ∪Ew j s(e,Aw p ,Aw q ) (3) In this direction, SEAT [13] proposes a generalization of WEAT to sentences.SEAT generates sentences generated by inserting words from WEAT's tests into simple templates such as "This is a[n] $word$."SEAT finds evidence of human-like bias in sentence encoders like WEAT does in word encoders.Specifically, a study conducted by [25] quantifies the degree to which gender bias differs with the corpora used for the pre-trained model and fine-tuning with additional data.Schroder et al. [26] proposes various metrics of embedding biases and compares their strengths and weaknesses.Garrido et al. [27] explore biases in embeddings through the lens of geometric spaces, providing a fresh perspective on the subject.Some studies [28,29] utilize embedding testing to investigate the evolution of biases across extensive temporal scales, revealing persistent statistical features throughout history.Furthermore, a recent study [30] reveals extensive potential for embedding testing.
Beyond uncovering biases and stereotypes, embeddings also provide precise assessments of the environments in which biases manifest within human culture and thought.
Although embedding-based approaches are related to IAT in connecting semantic distance and mental association, measured biases are a joint effect of explicit and implicit bias, where neither explicit nor implicit bias is correctly understood.For example, if people dislike flower in implicit bias but claim to like it in explicit bias, in their statements explicitly expressing their attitude toward flowers, concept "Flower" will be semantically closer to attribute "Pleasant" than "Unpleasant".However, in his/her rest sentences, "Flower" can be closer to "Unpleasant" than "Pleasant", which reveals their real implicit bias.Significantly, with the recent emergence of large language models (LLMs) extensively utilized in realworld applications [31,32], it is essential to differentiate between explicit and implicit biases for a thorough assessment of biases present in LLMs.

Automatic Sentiment Analysis
Lastly, we introduce the automatic sentiment analysis utilized for analyzing sentiment polarity.Sentiment analysis, otherwise known as attitude and opinion mining, refers to the use of natural-language processing, text analysis, computational linguistics, and biometrics to systematically identify, extract, quantify, and study affective states and subjective information [9].It aims to determine the attitude of a speaker, writer, or other subjects with respect to some topic or the overall contextual polarity or emotional reaction to a document, interaction, or event.Sentiment analysis involves classifying attitudes in text into categories like "positive", "negative", or "neutral" [33].Generally, current sentiment analyzers often represent an attitude with the following key factors [34]: Object: an entity which can be a person, event, product, organization, topic, etc.; Attribute: a feature of objects with respect to which evaluation is made; Attitude orientation or polarity: the orientation of an attitude on an object representing whether the attitude is positive, negative, or neutral; Attitude holder: the person or entity that expresses the attitude.The definitions of "object" and "attitude" in sentiment analyzers of attitude can match those of "concept" and "attitude" in psychological measurements.Although "attribute" has different definitions in automatic methods and psychological measurements, we do not need to match them in this work because we are not working on aspect-based attitude analysis [35].It is noted that state-of-the-art sentiment-based attitude measurements also do not clearly distinguish between explicit and implicit attitudes bias [36][37][38].In this study, we employ pre-trained machine-learning models [39] to analyze the sentiment conveyed in statements.Pre-trained machine-learning methods refer to the process of initially training a model on a large-scale dataset to acquire knowledge of the underlying language structure, followed by fine-tuning for specific tasks.Specifically, two sentiment analysis models, namely Stanford CoreNLP [40] and Baidu Sentiment Analyzer https://ai.baidu.com/tech/nlp_apply/sentiment_classify (accessed on 12 March 2024), were employed.These two models combine probabilistic models, machine learning, and linguistic knowledge to analyze sentiments of statements.
Although automated sentiment analysis has been widely studied and applied in fields such as public opinion analysis and human-machine dialogue, there has been no research yet on applying sentiment analysis techniques to identify people's biases.This not only requires us to theoretically attempt to correlate the key elements of sentiment analysis with the elements of bias measurement in the field of psychology but also necessitates the design of reasonable automated methods to apply this correlation to real-world data.

Highlight the Issues of Related Works
Here, we enumerate the key issues that need to be addressed in current related research and are also the main focus of this paper: (1) The SRA and IAT tests in psychology, which target explicit and implicit biases, require active cooperation from the subjects and are contact-based tests, making it difficult to accurately measure them in large-scale populations.(2) While the WEAT method based on word vector technology can automatically measure biases in large-scale populations, it is unable to distinguish between explicit and implicit biases.(3) Although automated sentiment analysis is widely used, it has not been applied to the automated measurement of explicit and implicit biases, and there is also a lack of analysis on the correlation between the two.

Our Approach
In this section, we will first introduce the high-level idea of the method proposed in this paper, followed by a detailed discussion of the key techniques involved.

Overview of the High-Level Idea of Connecting SRA/IAT and Linguistic Statements
Automatic measurement of attitude bias plays an essential role in many social computing studies, e.g., customized recommendation, public opinion analysis, influence analysis, and user profiling.Unlike the traditional psychological measurements in the laboratory, these tasks often require attitude measurement on a massive population instead of a limited sample set.Furthermore, different types of attitudes are expected for different applications.For example, in public opinion and influence analysis, people's explicit attitude bias is focused on, which is transferred in the group and influences each other through language.However, in customized recommendation and user profiling, people's implicit attitude bias can be more important, as it determines people's behavior inherently.
In psychology studies, subjects' explicit bias of the concepts can be measured when the subjects are consciously aware of the measured concepts, e.g., being asked questions containing the concepts in SRA.Subjects' implicit bias of the concepts can be measured when the concepts are implicitly primed in subjects' minds, but they are not consciously aware of it, e.g., giving responses to the exemplars instead of concepts in IAT [1].
For this task, both classical psychological measurements and current computational sentiment or semantic measures have certain defects.
First, in classical psychological measurements, although SRA and IAT can measure the explicit and implicit attitude separately with good theoretical validation, they have two main disadvantages for automatic massive measurements.

1.
Sample size: Due to SRA and IAT requiring active cooperation from the subjects, the size and distribution of the sample size are highly limited by how many and what kind of subjects can be invited.

2.
Independence: In both SRA and IAT, the subjects are aware that they are attending a measurement and may behave differently from their usual behaviors, i.e., the subjects cannot be exactly independent of the measurement.
Second, in current language-based computational measurements, although sentiment and semantic analysis can be applied to massive subjects, the main disadvantage is ignoring the difference between explicit and implicit bias.This will be confused in the case that the explicit and implicit biases are different.For example, if a group of people expresses positive explicit bias toward Black people, but they have a negative implicit bias in their mind, the sentiment or semantic scores on their language about Black people can be neutral in general.This result can mislead both the public influence analysis (explicit bias) and the user preference profiling (implicit bias).
To address the above problems, in this task, we propose to combine the advantages of psychological measurements and language-based computational measurements.Figure 3 illustrates the overview of the high-level idea of connecting SRA/IAT and linguistic statements.We simulate SRA and IAT with the sentiment of the statements containing combinations of concept/exemplar words (e.g., Flower/Rose) and attribute words (e.g., beautiful).In a set of statements expressing the attitude of the exemplars (e.g., Rose is beautiful.),speakers' implicit bias can be understood statistically because there is a chance for every statement that the concepts (e.g., Flower) implicitly influence speakers' attitudes toward the exemplars (e.g., Rose), but the speakers are not consciously aware of it.Furthermore, to connect the criteria of SRA/IAT to the linguistic features, we make the following hypothesis.In the statements expressing the attitude of the concepts, speakers' explicit bias can be understood because the speakers must be consciously aware of the concepts they are discussing.For example, in Figure 3, the statement "Flower is not beautiful at all" must express the explicit bias on the concept "Flower".However, in the statements expressing the attitude of the exemplars, speakers' implicit bias may be understood because there is a chance that the concepts implicitly influence speakers' attitudes toward the exemplars, but the speakers are not consciously aware of it.For example, in Figure 3, the statement "Rose is beautiful at all" may express the implicit bias on the concept "Flower".Although this implicit influence is not guaranteed in each statement of exemplars, it is still likely to understand this concept's implicit bias if we have many statements containing the exemplars of a concept.
Therefore, our high-level idea is to simulate the criteria of SRA or IAT by measuring the sentiment of people's statements on concepts or exemplars w.r.t attributes, thus distinguishing two kinds of statements involving explicit and implicit bias, respectively.In this way, we can understand people's explicit and implicit biases by calculating the sentiment scores of each type of statement.This simple idea has three advantages: (1) There is no need to modify state-of-the-art sentiment analyzers.
(2) The massive population statements are easy to obtain from online corpora.Furthermore, by narrowing the properties of the subjects or the corpus, it is possible to control the properties of the samples.For example, it is possible to find the bias of the female subjects or find people's bias in formal scenarios (from serious text like Wikipedia or news) or informal scenarios (from social media) (3) The proposed test is strictly independent of the data.The measurement of the bias in language will not impact the generation of the bias in language, which is a good feature for the measurement of psychological characteristics (neither SRA nor IAT exactly satisfies this independence).

Matching Psychological Tests, Types of Bias and Linguistic Statements
As shown in the left half of Figure 3 (the part covered by the indicator bar labeled "Matching Psychological Tests, Types of Bias and Linguistic Statements"), we propose two stages to simulate SRA/IAT linguistically: 'naming special words' and 'simulating SRA/IAT with statements of special words'.

Naming Special Words
As shown in the left half of Figure 3 (the part covered by the indicator bar labeled "Naming special words"), we define the special words of our approach.In SRA and IAT, there are three key factors: Concept, Exemplar and Attribute.We define three kinds of words naming these three key factors: (1) Concept words: concept words are the word naming concept, e.g., the concept words of concept "Flower" can be flower, flowers, . . .}. (2) Exemplar words: exemplar words are the words naming the exemplars of a concept, e.g., the exemplars words of concept "Flower" can be {aster, clover, hyacinth, . . .}. (3) Attribute words: attribute words are the words naming the connotation of an attribute, e.g., the attribute words of attribute "Pleasant" can be {happy, good, beautiful, . . .}.

Simulating SRA and IAT with Statements of Special Words
Then, as shown in the left half of Figure 3 (the part covered by the indicator bar labeled "Simulating SRA and IAT with statements of special words"), we simulate SRA and IAT with statements containing the combinations of special words.
(1) Explicit bias and statements of "concept word + attribute word": in SRA, subjects' biased expression on a concept with respect to an attribute is regarded as explicit bias on the concept.In this work, corresponding linguistic criteria is: if a statement mentions a concept word with respect to an attribute word, the sentiment of this statement is regarded as a case for explicit bias on the concept, e.g., the sentiment of "Flower is not beautiful" is a case of negative explicit bias on "Flower".(2) Implicit bias and statements of "exemplar word + attribute word": in IAT, the subject's biased response on an exemplar with respect to an attribute is regarded as implicit bias on the concept involving the exemplar.In this work, the corresponding linguistic criterion is: if a statement mentions an exemplar word with respect to an attribute word, the sentiment of this statement is regarded as a case of implicit bias on the concept involving the exemplar.For example, the sentiment of "Rose is beautiful" is a case of positive implicit bias on "Flower".
It is noted that we cannot only use the sentiment polarity of attribute words to determine the sentiment of statements.For example, in a statement, "Rose is not beautiful at all.", though the attribute word "beautiful" is positive in sentiment, the statement is negative in the implicit bias on "Flower".
In practice, the statements containing the "concept/exemplar word + attribute word" are identified according to grammatical dependency between the concept/exemplar word and attribute word.In particular, a statement is selected if and only if it is a "subject-verb-object" or "subject-link verb-predicative" where a concept/exemplar word and an attribute word are both involved.

Calculating Explicit and Implicit Bias with Sentiment Scores
As shown in the right half of Figure 3 (the part covered by the indicator bar labeled "Calculating Explicit and Implicit Bias with Sentiment Scores"), matching the two categories of statements with SRA and IAT, we separately calculate the sentiment scores of the collection of each category of statements to measure the explicit and implicit bias.
Given a pair of concepts C i vs. C j , Cw i and Cw j are the concept word sets of C i and C j , respectively.Ew i and Ew j are the exemplar words sets of C i and C j , respectively.Aw p and Aw q are the attribute words sets of a pair of attribute A p vs. A q , respectively.Given statements set S, the following steps calculate the explicit and implicit bias between C i and C j of the people who generate S.
S Cw_i , S Cw_j , S Ew_i , S Ew_j ⊆ S are four statements collections in which each statement simultaneously contains one word from Cw i , Cw j , Ew i , Ew j (concept or exemplar words), and one word from Aw p ∪ Aw q (attribute words), respectively.S Cw_i , S Cw_j , S Ew_i and S Ew_j are used to identify the explicit bias on C i , the explicit bias on C j , the implicit bias on C i and the implicit bias on C j , respectively.Equations ( 4) and ( 5) calculate the explicit and implicit bias on C i vs. C j with respect to A p vs. A q , respectively, where p(s) is the sentiment score of statement s.

Experiments
In this section, we will present the experiments and analysis conducted in this paper, including the grouping and objectives of the experiments, the experimental design, the results obtained, and a discussion of the findings.

Aims of Experiments
In the experiments, we investigated two questions: The first question is: applying our method on corpora, can we reproduce well-known psychological observations (ground truth) of explicit and implicit bias?
The second question is: with our method, can we have new findings besides the well-known observations?
Then, we had three groups of experiments: The first experiment is: on various online corpora, comparing the experimental observations of our proposed method with those reported by SRA and IAT (Section 4.3).
The second experiment is: examining the effects of the selection of special words and sentiment analyzers (Section 4.4).The concept of sentiment analyzers is introduced in Section 2.3.
The third experiment is: comparing the evolution of explicit and implicit bias over time (Section 4.5).

Global Settings
We introduce the basic settings used in our experiments, including the selection of different corpora, the target words of the study, the sentiment analysis tools utilized, and the evaluation methods.These settings allow us to systematically investigate and compare explicit and implicit biases across different languages, media, and cultural contexts.

Corpora
Four corpora of different languages and media were explored.For serious corpora, English Wiki (99 M sentences) and Chinese Wiki (7 M sentences); for social media corpora, Twitter (100 M English sentences) and Weibo (77 M Chinese sentences).For each concept and each corpus, we selected two subsets of two kinds of statements defined in Section 3.2.2,i.e., (1) subsets of the statements of "concept word + attribute word"; (2) subsets of the statements of "exemplar word + attribute word".Stanford Dependency Parser [41] was used to identify the "subject-verb-object" or "subject-link verb-predicative" pattern in statements.

Concept Pairs and Special Words
In experiments, people's bias on four concept pairs was investigated.To compare our results to well-known psychological observations, four concept pairs reported in IAT were selected including "Insect vs. Flower", "Weapon vs. Instrument", "Afri-American vs. Euro-American" and "China vs. America".The selected concepts involve general social bias, controversial social bias, and cultural differences between China and America to make the experiments more representative.
In English corpora, concept words were selected to name the concept directly.The exemplar words and attribute (pleasant and unpleasant) words were selected following IAT [5], WEAT [12], and Chinese version of Harvard's online IAT https://implicit.harvard.edu/implicit/china/ (accessed on 12 March 2024).We manually translated the special words used in English experiments into Chinese in Chinese experiments.
The detailed list of selected special words for each pair of concepts is as follows: ( In experiments, sentiment scores for English and Chinese statements were obtained using two sentiment analysis tools, referred to as Stanford CoreNLP and Baidu Sentiment Analyzer, as mentioned in Section 2.3.The reason for choosing these two sentiment analyzers is that they are among the best-performing sentiment analyzers in English and Chinese, respectively.To verify the performance and compatibility of these two analyzers, we constructed 100 Chinese-English test statement pairs (listed in Appendix A.2), where each pair consists of semantically identical English and Chinese statements containing both concept/exemplar words and attribute words.We analyzed the sentiment polarity of English and Chinese statements in each pair using CoreNLP (https://stanfordnlp.github.io/CoreNLP/, accessed on 12 March 2024) and Baidu Analyzer (https://intl.cloud.baidu.com/product/body.html,accessed on 12 March 2024), respectively.The results showed that CoreNLP achieved a 94% accuracy rate on 100 English statements, while Baidu Analyzer achieved a 97% accuracy rate on 100 Chinese statements, with a 96% matching degree between the two analyzers.This indicates good consistency between their sentiment analysis results.
To examine the effects of different sentiment analyzers, in the testing of effects of procedural variables, we chose two comparative analyzers: NLTK sentiment package for English and Tencent Sentiment Analyzer https://ai.qq.com/doc/nlpemo.shtml(accessed on 12 March 2024) for Chinese.

Evaluation Method
To the best of our knowledge, our work is the first to automatically distinguish explicit and implicit bias in corpora.We build a high-level and intuitive connection between SRA/IAT and linguistic statements.Although the validity of such a connection cannot be theoretically guaranteed, we propose to use well-known psychological observations [5] as ground truth and investigate whether the proposed method can reproduce these observations qualitatively.The observations will be compared in Section 4.3.In Section 4.3, we employed sentiment scores as evaluation metrics to compare explicit and implicit biases across different datasets and concept pairs, revealing disparities between explicit and implicit biases under various conditions.Furthermore, in Section 4.5, we utilized the temporal changes in sentiment scores to compare the trends of explicit and implicit biases across different datasets and concept pairs, highlighting the differences in their dynamic evolutionary patterns, which have not been explored in traditional psychological research.

Comparing Explicit and Implicit Bias
Results: In the first group of experiments, we investigated the performance of our method of distinguishing explicit and implicit bias.We calculated the explicit and implicit bias separately with our method on each corpus for each concept pair.The significance of the difference between explicit and implicit bias pairs was tested with a permutation test using no difference as the null hypothesis.The results are shown in Figure 4, where the explicit and implicit bias of each concept pair on each corpus is compared using sentiment scores calculated with Equations ( 4) and ( 5).Positive and negative values indicate the bias preferring the right and left concept of this column, respectively (e.g., in the first column, Insect is the left concept, Flower is the right concept).The red star indicates that the difference between this pair of explicit/implicit biases is significant (p = 0.0001)."Afri-A" = "Afri-American", "Euro-A" = "Euro-American".For example, in English Wiki, people prefer to Afri-American in explicit bias and prefer to Euro-American in implicit bias.Discussion: In Figure 4, we have four main observations: (1) In general: implicit bias can be significantly different from explicit bias, which agrees with the IAT observation.The difference is sensitive to the sociality of concepts, scenarios, and cultures, which agrees with the psychological assumptions of implicit cognition: the difference between explicit and implicit bias can be explained by the social desire [5].(2) On concepts: in most cases, the difference between implicit and explicit bias is more significant on social and controversial concept pairs (i.e., "Afri-American vs. Euro-American" and "China vs. America") than the difference on unsocial concept pairs (i.e., "Insect vs. Flower" and "Weapon vs. Instrument").This observation agrees with the psychological assumption that for social and controversial concepts, people are more motivated to adjust their explicit bias to adapt to social desire, which leads to greater differences between implicit and explicit bias.(3) On scenario: for controversial concept pairs, implicit and explicit biases are closer to each other in social media than in Wikipedia.From social desire, users are more strongly required to conform to mainstream social values in Wikipedia than in social media.
(4) On language (culture): though on both two English corpora, explicit and implicit bias on "Afri-American vs. Euro-American" have opposite polarity (consistent with IAT reports), but this does not happen in Chinese corpora.Similarly, explicit and implicit bias on "China vs. America" are opposite on Chinese corpora instead of on English corpora.From social desire, the importance of consistency with mainstream social values can vary across countries or cultures for a concept pair.
From the view of the relationship with previous work, our experimental results line with psychological observations in general.Our innovation lies in the automation of explicit and implicit bias analysis by correlating concepts from psychological measurements with elements of automated sentiment analysis.Furthermore, we addressed the shortcomings of the previous word vector-based automated analysis method, which was unable to distinguish between explicit and implicit biases.The limitation of our approach is that further explanation of the difference in specific concept pairs still needs help from psychology scientists in future work.This explanation might achieve new findings considering the absence of observations on a massive population in the current psychological study on implicit bias.

Effects of Procedural Variables
Besides p-values of the difference between explicit and implicit bias, we also tested the significance of the qualitative observations in Section 4.3 by testing the effects of procedural variables.The proposed method has two procedural variables: special words and sentiment analyzers' selection, and this section investigates whether these two variables affect the experimental observations.Table 2 illustrates the percentage that experimental observations are maintained across procedural variables, where 'Words' is special words, 'Analyzer' is sentiment analyzers, 'E-' is English, and 'C-' is Chinese.
For the effect of exemplar word selection, we randomly selected a subset from the original exemplar word set and repeated the experiments.We repeated the experiment in Section 4.3 1000 times and calculated the percentage of the qualitatively consistent observations in Section 4.3.
For the effect of sentiment analyzer choice, we randomly selected a subset from the original exemplar words set, repeated the experiments on the subset with comparative analyzers indicated in Section 4.2.3, and compared the original analyzers' results.We repeated the test 1000 times and calculated the percentage that the results of comparative and original analyzers were consistent with each other.
In Table 2, the observations drawn in Section 4.3 are maintained at high probability across different exemplar words and sentiment analyzers.The concept of sentiment analyzers is introduced in Section 2.3.

Evolution of Explicit and Implicit Bias
Results: In this group of experiments, we compare the evolution of explicit and implicit bias over time.We use English and Chinese social media corpora (i.e., Twitter and Weibo), with a timestamp for each statement.To track the evolution of the biases, we broke each corpus into subsets of months, spreading over 36 months.Then, we measured explicit and implicit bias month by month using the proposed sentiment-based bias measurement.In Figures 5 and 6, we illustrate the detailed evolution tracks of each concept pair on each corpus.In Figure 7, we compare the stability of explicit and implicit bias in evolution.The stability is calculated with the standard deviation of the bias over time.
Discussion: In Figures 5-7, we have two main observations: (1) Explicit and implicit bias is different in stability: beyond existing psychological studies, a new observation of evolution experiments is that implicit biases are more stable than explicit biases in most cases.In Figures 5-7, with the exception of the concept pair "China vs. America", implicit biases in other concept pairs are more stable than explicit biases.This may indicate that although people's explicit public bias may vary due to the changing social context (e.g., the scenario or the talking partner), their internal implicit bias tends to be more consistent over time.This possibly explains the finding that implicit bias is a better predictor of certain kinds of behavior than explicit bias [6,7].(2) Factors influencing the stability: in this experiment, the sociality of concepts also influences the evolution of biases.In Figures 5 and 6, implicit biases of unsocial concept pairs (i.e., "Insect vs. Flower" and "Weapon vs. Instrument") are more stable than that of social concept pairs (i.e., "Afri-American vs. Euro-American" and "China vs. America").
In terms of its relationship with previous work, the dynamic analysis presented in this paper builds upon the existing concepts of psychological bias measurement.However, its novelty lies in the fact that existing psychological measurements primarily focus on static analysis, whereas our study delves into the dynamic evolution and comparison of explicit and implicit biases.While our work has successfully revealed the differences in the evolution of these two types of biases, its limitation lies in the fact that we have not yet empirically and deeply analyzed the impact of these differences on behavior.This aspect remains an area for further exploration in future research.

Comparing WEAT and SEAT with Proposed Sentiment-Based Method
As introduced in Section 2.2, WEAT [12] and SEAT [13] are a joint effect of explicit and implicit bias.The factors of our method are more naturally connected with those of SRA and IAT, and it can distinguish explicit and implicit bias.Experimental results support the effectiveness of this connection.In the case that explicit and implicit bias might be quite different, and the difference is seriously concerned with the task, our method can be a better choice for bias identification.Furthermore, due to the requirement of massive data in high-quality training of word or sentence embedding, WEAT and SEAT do not work well on small data (e.g., the language of an individual).However, our method derives bias cases from every single statement and is not as sensitive to the data size as WEAT does.

The Application of This Work in Eliminating Bias in Generative AI
Bias is also a key issue in current research on generative AI.There is research indicating that the content generated by generative AI may contain bias [42][43][44].This bias is considered inappropriate and may affect the user's cognition.Although there have been some attempts to identify and eliminate biases in generative AI, an important drawback of current efforts is the lack of clear differentiation between explicit and implicit biases.This drawback makes it difficult to accurately evaluate the cognitive characteristics of bias in generative AI, therefore reducing the effectiveness of bias elimination, especially implicit bias elimination.Our latest research shows that biases in generative AI can also be distinguished as explicit and implicit [45].Therefore, the work of this article will help automate the identification and differentiation of explicit and implicit bias in generative AI, thus more effectively guiding research on bias elimination in generative AI.

Summary
Using automatically obtained sentiment scores of statements containing the combination of special words on various corpora, we automatically reproduce psychological measurements of explicit and implicit attitude bias in a large population.We examine the special words of four pairs of concepts and the statements from four large corpora, including English and Chinese Wikipedia and social media.The experimental results reveal that our automatic method qualitatively agrees with the costly small-scale observations from psychological studies.Specifically, explicit and implicit attitude bias can be significantly different.The significance of the difference is shown to be influenced by the sociality of concepts, the scenario of language use, and the culture behind the language.We also find that implicit attitude biases are more stable than explicit attitude biases over time, which is absent in existing psychological studies.The theoretical significance of our work is to provide a new perspective based on sentiment analysis for the concept of implicit association in traditional psychological research.The technical significance is in proposing an automatic measurement method that can replace high-cost traditional psychological measurements.The practical significance is in providing new research tools for explicit and implicit cognition targeting larger-scale populations.

Future Work
In future work, there are two main directions to further improve this work.The first direction is to examine the conclusions of this work with more datasets.Although the corpora used in this work cover two main languages (English and Chinese) and the two most common forms of Internet text (serious text from Wikipedia and free text from social media), the experimental results of data from more different languages and sources are expected to provide more evidence for the conclusions of this work and help to understand the explicit and implicit bias of people from different cultural backgrounds.Another important future work is to design algorithms to automatically discover the concepts on which people have bias instead of selecting the concepts manually.Our work in this paper manually adopts the specified concepts to facilitate comparison with existing research in psychology to verify our method.Once the effectiveness of our method is verified, in practical applications, we need to be able to automatically discover more biased concepts that have not been addressed in existing research by improving our method.
Implicit bias: implicit bias is one's internal bias without conscious awareness or honestly reporting [47].An individual's consciousness does not control implicit bias.The primary purpose of measuring implicit bias is to predict behaviors that explicitly held biases cannot predict [49].For example, although one deliberately claims to prefer African Americans to European Americans, he/she may associate African Americans with negativity in implicit cognition [5].
Concept: concepts name the objects on which people have bias [47].As psychological tests do, we pair concepts to understand bias, e.g., Afri-American vs. Euro-American.
Exemplar: exemplars are stimuli of concepts.Exemplars should be objects which can be clearly classified into one concept in a concepts pair, e.g., typical African-American names or photos can be the exemplars of concept "Afri-American" w.r.t the concepts pair "Afri-American vs. Euro-American" [47].
Attribute: attribute is the connotation of bias.For example, in IAT, attributes of positive and negative bias can be "Pleasant" and "Unpleasant", respectively [12,47].
Self-Report Assessment (SRA): Self-report assessment is a typical explicit bias method.Reading questions and choosing responses in the assessment is the most direct method of asking about one's feelings, prejudices, beliefs, and other information [17].SRA can be a survey, questionnaire, or poll in which subjects express their attitude explicitly.The main strength of SRA is allowing participants to describe their own experiences rather than inferring this from observing them.
Implicit Association Test (IAT): IAT [5] is the most widely used psychological test of implicit bias.IAT is designed to measure the strength of subjects' implicit associations between a pair of concepts (e.g., Flower vs. Insect) and a pair of attributes (e.g., Pleasant vs. Unpleasant).

Appendix A.2. Sentence Pairs for Comparing the Two Sentiment Analyzers
In this section, we enumerate all sentences employed to compare the consistency of sentiment analyzers between Stanford CoreNLP and Baidu Sentiment Analyzer.The sentences comprise two groups, positive and negative, with 50 pairs in each group.Each pair includes an English sentence and the corresponding semantically equivalent Chinese sentence.
The positive sentiment sentences are enumerated as follows: The gentle melody of the violin brings peace to my soul; 小提琴的温柔旋律让我的灵 魂感到平静.
Roses add beauty and love to every garden they grace; 玫瑰花为每一个花园增添了美 丽与爱.
The lilac's fragrance is a miracle of nature, evoking pure happiness; 丁香花的香气是 大自然的奇迹, 唤起纯粹的幸福.
A family picnic in the park, surrounded by blooming daisies, feels like heaven; 在公园 里举行家庭野餐, 周围布满开放的雏菊, 感觉如同天堂.
The laughter at the concert was a testament to the harmonica's cheerful sound; 音乐会 上的笑声证明了口琴的愉快声音.
A rainbow appeared over the field just as we spotted a beautiful bluebell; 当我们看到 美丽的风铃草时, 天空中出现了一道彩虹.
The honest lyricism of the guitar struck a chord with everyone in the room; 吉他的诚 实歌词与在场的每个人产生了共鸣.
The breeze through the magnolia flowers gently caressed me, which was really peaceful; 穿过木兰花的微风轻抚, 真是太平静了.
Playing the piano at sunrise is a magical experience, filled with pleasure; 在日出时弹 奏钢琴是一种神奇的体验, 充满了愉悦.
Learning to play the flute has been a journey filled with joy and friendship; 学习吹长 笛的过程充满了欢乐和友谊.
The sight of a butterfly on a zinnia is a small, everyday miracle; 蝴蝶停留在紫罗兰上, 这是一个微小的日常奇迹.
The love song played on the saxophone was pure romance; 萨克斯管演奏的恋歌纯粹 是浪漫.
Spending a vacation surrounded by wildflowers is truly a slice of heaven on earth; 在 野花丛中度假是人间天堂.
The health benefits of walking among orchids are as pleasant as their beauty; 在兰花 中散步的健康益处与它们的美丽一样令人愉悦.
The tulip festival was a rainbow of colors and a feast for the senses; 郁金香节是色彩的 彩虹, 是感官的盛宴.
The early-blooming crocuses, fortunately, discovered the joy that heralds the arrival of spring; 早开的番红花幸运地发现, 预示着春天的欢乐.
The harp's melody at the wedding was a gift of heavenly sound; 婚礼上的竖琴旋律是 天籁之音.
The family gathers annually to enjoy the cherry blossoms, a tradition of happiness; 家 人每年聚集一起欣赏樱花, 这是幸福的传统.
A gentle dragonfly rests on a lily, a peaceful scene in nature's playground; 一只温柔的 蜻蜓停在百合花上, 这是大自然游乐场中的一个和平场景.
The peony garden was a place of peace and gentle beauty; 牡丹花园是和平与温柔美 丽的地方.
The trumpet's blast at the parade spread cheer among the crowd; 游行中小号的声音 在人群中传播欢乐.
The pleasure of watching hummingbirds dart among the azaleas is unmatched; 观看 蜂鸟在杜鹃花间穿梭的乐趣无与伦比.
The violinist's performance was a diamond sparkling in the evening; 小提琴手的表演 如同晚上的一颗闪耀的钻石.
The laughter of children chasing butterflies is pure happiness; 孩子们追逐蝴蝶的笑声 是纯粹的幸福.
A rose in full bloom is a symbol of love and natural beauty; 盛开的玫瑰是自然美和爱 的象征.
The clarinet's soft notes added a layer of peace to the quiet afternoon; 单簧管的柔和音 符为安静的午后增添了一层平和.
The paradise of a spring garden is enhanced by the scent of hyacinths; 春日花园的天 堂由风信子的香味增强.
The loyal old dog peacefully lies among the buttercups in the sunshine; 忠诚的年老的 狗在阳光中平和地躺在金凤花中.
The enchanting melodies of the mandolin near the campfire create a stronger sense of camaraderie among friends; 营火旁的曼陀林声音使朋友们更亲近.
The miracle of nature is evident in every unfolding daffodil; 每一朵绽放的水仙花都显 现自然的奇迹.
The marigold's bright colors bring happiness to any sunny spot; 金盏花的鲜艳色彩为 任何阳光充足的地方带来快乐.
The lucky discovery of a rare orchid made the hike unforgettable; 偶然发现一朵罕见 的兰花使徒步之旅难忘.
The pleasure of a trombone's slide brings jazz to life in the city; 滑音大号的声音让城 市的爵士乐生动起来.
A sunny vacation day is best spent surrounded by blooming violets; 阳光明媚的假日 最适合被盛开的紫罗兰环绕.
The laughter of friends around the drum circle creates a cheerful atmosphere; 鼓圈周 围朋友的笑声营造出愉快的氛围.
The diploma hung proudly above the piano, a symbol of achievement and music; 毕 业证书骄傲地挂在钢琴上方, 是成就与音乐的象征.
The gentle breeze brushing against the iris bed feels soothing; 微风轻抚鸢尾花床, 感 觉很舒缓.

Figure 1 .
Figure 1.A typical interface of computer-aided IAT on the implicit bias of African-American vs. European-American.

Figure 2 .
Figure 2. A schematic diagram of a real scene where a subject is taking an IAT test in front of a computer.

Figure 3 .
Figure 3. Overview of the high-level idea of connecting SRA/IAT and linguistic statements.

Figure 4 .
Figure 4. Compare explicit and implicit bias of each concept pair on each corpus.

Figure 5 .
Figure 5.The evolution of explicit and implicit bias on Twitter.

Figure 6 .Figure 7 .
Figure 6.The evolution of explicit and implicit bias on Weibo.

analysis Attitude Bias Psychological measurements Statements of special word combinations Simulate Explicit bias Implicit bias Attitude Bias Explicit bias Implicit bias
Implicit Association Test (IAT) on exemplars and attributes Self-Report Assessment (SRA) on concepts and attributes Exemplar Statements on exemplar words of "exemplar word + attribute word" (e.g., "Rose is beautiful.")Corpus Statements on concept words of "concept word + attribute word" (e.g., "Flower is not beautiful at all.")Average sentiment score Sentiment Psychological measurements based on subjects' response on stimuli Our automatic measurements based on people's language Matching

Table 1 .
Count of explicit and implicit statements in experiments.

Table 2 .
The percentage that experimental observations are maintained across procedural variables.