Next Article in Journal
New Space Engineering Design: Characterization of Key Drivers
Previous Article in Journal
Combining QSAR and Molecular Docking for the Methodological Design of Novel Radiotracers Targeting Parkinson’s Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generative Language Models for Personality-Based Utterances in Novels: A Character Clustering Approach

by
Eun-Jin Kim
1,
Chung-Hwan Joe
2,
Misun Yun
3 and
Young-Seob Jeong
1,*
1
Department of Computer Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea
2
Department of English Language and Literature, Hongik University, Seoul 04066, Republic of Korea
3
Division of English and European Studies, Inha University, Incheon 22209, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(15), 8136; https://doi.org/10.3390/app15158136
Submission received: 11 May 2025 / Revised: 4 July 2025 / Accepted: 18 July 2025 / Published: 22 July 2025

Abstract

In novels, readers encounter a variety of characters with distinct personalities, and their satisfaction tends to increase when each character’s utterances consistently reflect their unique traits. Recent advances in large language model (LLM) technology have made it possible to perform complex tasks such as generating long-form narratives and adapting writing styles. However, research on generating character utterances that reflect individual personalities remains limited. In this paper, we identify a key challenge in this task, namely the unconscious influence of the author’s writing style, and propose a novel clustering-based method to mitigate this problem by tuning large language models. We manually annotated Big Five personality trait scores for characters appearing in selected novels and designed prompts to generate examples for instruction-tuning. Experimental results demonstrate that language models trained using our proposed method produce utterances that more consistently reflect character personalities compared to untuned models.

1. Introduction

Large language models are pre-trained on vast amounts of data and are utilized for various tasks such as machine translation, sentiment analysis, and document summarization. Generative language models, such as GPT-3 [1] and LLaMA [2], in particular, have played a central role in the field of natural language processing and have made significant progress in recent years. However, despite the rapid progress of LLMs, their general-purpose nature can lead to suboptimal performance compared to models specialized for specific domains or tasks (e.g., code generation [3], medical applications [4,5], and style-based literature writing [6]).
While current research on novel generation using LLMs primarily focuses on maintaining consistent story plots [7]. In contrast, relatively little research has focused on the consistency of character presentation. Existing approaches to character-centered storytelling focus only on aligning character actions with consistent character presentation [8], and do not consider personality-level consistency in character utterances.
When reading a novel, if the characters speak consistently with their personalities, it enhances the sense of realism; otherwise, the reader’s satisfaction may decrease. Writers strive to create realistic characters for their readers, but often characters with different personalities speak in similar ways. Prior studies have shown that an individual’s linguistic style (idiolect) tends to be consistently reflected in their writing [9], and that an author’s stylistic tendencies appear throughout a text, whether consciously or unconsciously [10]. It has also been reported that an author’s stylistic tendencies may unintentionally manifest in character utterances, potentially weakening the unique characteristics of the characters [11]. We call this “unconscious author style,” that the writer unconsciously influences the character’s thoughts, emotions, and utterances [12,13]. That is, “unconscious author style” refers to the phenomenon in which an author’s linguistic style unintentionally intrude upon the utterances of characters, thereby blurring or diluting their personalities. Minimizing such interference would facilitate the generation of utterances that align more closely with each character’s personality traits, resulting in more plausible and consistent dialogue.
As far as we know, there have been no studies aimed at reducing the “unconscious author style” on characters’ utterances. Some previous studies have tried to generate character utterances, but they were only for free conversation, drama or games, without considering the “unconscious author style” issue [14,15]. In this paper, we focus on the way of speaking (i.e., linguistic style) of characters in novels. Specifically, based on previous findings that the linguistic style has a strong relationship with personality [16], we design a new method that generates personality-consistent utterances of characters in novels. To achieve this goal, we use instruction-tuning to construct a new language model that is tuned to generate utterances that reflect the personalities of the target characters, while suppressing the influence of “unconscious author style” on their dialogue.
To handle the personality of the characters, we adopt the Five-Factor Model of Personality (FFM), also known as the “Big Five” or OCEAN traits: Openness, Conscientiousness, Extraversion, Agreeableness, and Neuroticism [17]. Based on the theory of OCEAN traits, we assume that the characters’ utterances reflect their distinct personalities even though they are not real people. To validate this, we manually annotated the OCEAN trait scores of novel characters, and generated two different character clusters: word-based clusters and personality-based clusters. The word-based clusters are obtained from the words of the characters’ utterances, while the personality-based clusters are created from the characters’ OCEAN trait scores. We observed that the word-based clusters show a similar trend to the personality-based clusters, i.e., the characters with similar personality traits have similar word distributions in their utterances.
Based on the finding of the above assumption, our hypothesis is that a language model trained on utterances of character clusters can translate a given utterance into a new utterance that better reflects the personality of a target character cluster. That is, even though a character’s utterances may reflect the “unconscious author style” (i.e., linguistic patterns of the author), using the utterances of numerous similar characters from different novels will reduce the impact of the “unconscious author style.” This study does not aim to prevent changes in a character’s personality, but rather focuses on addressing the problem of the unconscious author style unintentionally influencing character utterances. If this hypothesis holds, then the language model will contribute greatly to the novel-writing process.
Suppose that a novelist has defined the personality of a main character in her/his novel, and she/he has found that the personality of the main character is close to the character cluster ‘A’. When the main character needs to deliver a line like “I never want to go back to my hometown,” the language model is prompted to generate a translated version of the line that matches the personality of the cluster ‘A’, so that the translated utterance will better reflect the personality that the novelist intended.
Figure 1 shows an example of such a translation, where the left conversation is adapted from the Sherlock Holmes stories “The Adventure of the Blue Carbuncle” [18] and “A Scandal in Bohemia” [19] and the right conversation is a translated version using our language model. The bottom of Figure 1 shows the personality difference between the characters by representative per-trait keywords [20]. Note that we have two target character clusters in the right conversation, each represented by Mary Crawford (Mansfield Park) and Gabriel Syme (The Man Who Was Thursday). The translated version is created by prompting our language model to translate the utterances of Holmes and Watson into the utterances of Crawford and Syme, respectively. As shown in the translated conversation, the translated utterances seem to successfully reflect the personality of Crawford and Syme.
This study addresses the following research question: Can we generate character utterances that better reflect their personality by using language models to mitigate the influence of the unconscious author style? Our contributions can be summarized as follows.
1.
We present a language model that is tuned for translating a given utterance into a new utterance that better reflects a desired personality in novel. As far as we know, this is the first language model designated for generating utterances reflecting personality in novels.
2.
We propose a method for building our language model. It involves (1) clustering characters using utterance words, and (2) prompt design for instruction-tuning data preparation using the character clusters.

2. Related Work

2.1. Transformer Decoder-Based Language Models

Since Transformer, many Transformer decoder-based language models have appeared and demonstrated their superior performance in natural language generation (NLG) tasks [21]. Previous studies have found that larger decoder-based models achieve better performance in many tasks, and such large language models (LLMs) can be divided into two groups: open-source LLMs and closed-source LLMs. The closed-source LLMs do not disclose their specification and parameter weights, and representative models include the GPT series [1,22,23] and Gemini [24]. On the other hand, the open-source LLMs provide publicly available source code, allowing community collaboration, customization, and transparency. Some models even provide training datasets. Representative open-source LLMs include the Llama series [2,25], Falcon [26], Gemma series [27,28], Pythia [29], and Mistral [30].
Because of their greater accessibility, the open-source LLMs are more widely used as backbone models for research purposes. Despite recent advances in training methods for decoder-based LLMs, domain- and task-specific models often outperform general-purpose models. This is because decoder-based LLMs are typically trained to be broadly applicable, and as a result, they tend to exhibit weaknesses in some domains. For example, as highlighted in a recent study by Gu et al. (2023), LLMs trained on general-purpose datasets often struggle to handle domain-specific elements such as frequently used APIs and libraries [3]. A representative example of a domain-specific model is Med-PaLM 2, which shows significantly better performance in the medical domain compared to models such as GPT-3.5 and PaLM [5,31]. Furthermore, existing studies on character utterance generation have focused solely on reflecting the personality of the character, without considering the issue of “unconscious author style.” There are several ways to tune the decoder-based LLMs. Few-shot learning demonstrates strong performance with only a small number of examples by leveraging their pre-trained knowledge. In particular, instruction-tuning is an effective prompt-based approach that trains models on explicit instruction-response pairs, enabling them to understand and execute complex and abstract commands. This method allows the model to be applied to different downstream tasks without requiring full parameter fine-tuning, while still achieving strong generalization performance. Furthermore, it is a parameter-efficient strategy that is well-suited for decoder-based LLMs, as it can yield competitive results even with limited training data. Thus, we construct a specialized LLM through the instruction-tuning that mitigates the influence of the “unconscious author style” while generating sentences that reflect the character’s personality.

2.2. The Big Five Personality Traits (OCEAN)

The Five Factor Model of Personality (FFM), also known as the “Big Five,” is a widely used theoretical model of human personality in recent psychology. It includes the OCEAN traits: Openness, Conscientiousness, Extraversion, Agreeableness, and Neuroticism. These traits are known to include various personality-related factors such as happiness, health, religion, identity, job satisfaction, political beliefs, etc. [32]. Each trait can be represented as a score (e.g., an integer value between −3 and 3), where the magnitude of the value indicates the strength of the corresponding trait. In addition, the traits can be represented by personality-related keywords, as described in Table 1 [20]. In this paper, we manually annotated OCEAN trait scores for characters from selected classic English novels and used them to test a correlation between character utterances and their personality. Details about the dataset annotation process and result are provided in the next section.

3. Materials and Methods

3.1. Validity of Using Character Utterances

Before developing our language model for utterance translation, we investigated whether the use of utterances really reflects the personality of characters. That is, although the previous studies have already found that linguistic style in novels has a strong relationship with personality [16], we wanted to answer the question “Do the characters who are similar in terms of personality really talk in a similar way?” because the characters are not real people, but only fictional characters.
Based on the Big Five traits, we first annotated the OCEAN scores of 28 novels from the Project Dialogism Novel Corpus (PDNC) [33]. PDNC was chosen for the data because it is a structured corpus of character utterances extracted from a variety of classic English and American novels, and because it is manually identified and annotated by experts in Western literature. Specifically, to control for factors such as genre and historical context, we used the PDNC dataset, which focuses on 19th- and early 20th-century English and American novels. It primarily includes works by authors like Jane Austen, E. M. Forster, and Henry James, whose writings center on domestic life and individual relationships, enabling experiments on a relatively cohesive and thematically consistent set of classic literary texts. The 28 novels include A Handful of Dust, A Passage to India, A Room with a View, Alice’s Adventures in Wonderland, Anne of Green Gables, Daisy Miller, Emma, Hard Times, Howards End, Mansfield Park, Night and Day, Northanger Abbey, Oliver Twist, Persuasion, Pride and Prejudice, Sense and Sensibility, The Age of Innocence, The Awakening, The Gambler, The Invisible Man, The Man Who Was Thursday, The Mysterious Affair at Styles, The Picture of Dorian Gray, The Sign of the Four, The Sport of the Gods, The Sun Also Rises, Where Angels Fear to Tread and Winnie-the-Pooh.
Each character in the novels is annotated with five scores ranging from −3 to 3, each corresponding to one of the five traits O, C, E, A, and N. The annotation was performed by five graduate assistants trained in literary studies. Each character was annotated by two annotators. Personality can be assessed through either self-report or report-by-others methods. Since the process of constructing a personality annotated dataset is inherently challenging, a recent study that released a new dataset for over 1000 individuals also adopted the report-by-others method for personality annotation [34]. We adopted the report-by-others method in this study, as the subjects of personality measurement are fictional rather than real individuals. We filtered characters that did not have enough utterances and obtained 139 characters. Pearson correlation coefficients are 0.51, 0.43, 0.30, 0.51, and 0.36, for O, C, E, A, and N.
According to [35], these coefficients are regarded as medium or large. We also computed Intraclass Correlation Coefficient (ICC) to assess the reliability of the scores annotated by the annotators for O, C, E, A, and N. ICC(3,1) [36], which measures the consistency of ratings for a single measurement, yielded the following values in our study: O = 0.50, C = 0.42, E = 0.31, A = 0.51, and N = 0.35. These generally showed fair levels of agreement (0.42) [37] with lower consistency observed for E and N. On the other hand, ICC(3,2), which assesses the reliability of the average of two raters, showed generally higher values in our study: O = 0.66, C = 0.59, E = 0.47, A = 0.67, and N = 0.52. These results suggest that the personality trait scores used in this paper provide a reasonable level of reliability. Examples of the annotated characters are shown in Table 2. The OCEAN distributions are shown in Figure 2.
Based on the annotated data, we created two different clusters: personality-based clusters and word-based clusters. Since the degree to which a character’s personality is expressed in their utterances may be limited and such utterances may also be influenced by the author’s stylistic tendencies, we cluster utterances from similar characters appearing in different works by different authors, in order to extract more representative personality traits. For the word-based clusters, we removed all nouns and digits in the utterances and performed stemming using a Porter stemmer. The characters with utterances shorter than 100 letters are filtered out. The word-based clusters are created by a bag-of-words approach with TF-IDF on 8000 vocabulary (i.e., unique words), using the k-means algorithm, and the number of clusters is 2. We also get average personality trait scores and normalize the scores to a range between 0 and 1, and create the personality-based clusters using the k-means algorithm. If the personalities of the characters are well reflected in their utterances (i.e., words), those with similar personalities should show close linguistic patterns to each other. Figure 3 shows the t-SNE results of the two different clusters, where the dots are characters. Characters with similar linguistic patterns are shown as nearby dots in the figure, and the color of each dot indicates its arbitrary personality cluster; we do not specify the personalities of the orange and blue clusters here, because we intend to check if the personality-based clusters are consistent with the word-based clusters. The figure shows that the personality clusters (i.e., the orange and blue clusters) have their own distinct centroids even though they overlap somewhat, which means that the characters with similar linguistic patterns tend to have the same personality. Based on this finding, we designed our method for tuning a language model for utterance translation. In particular, we expected that using utterances from similar characters (i.e., characters in close proximity in Figure 3) would reduce the impact of the “unconscious author style” problem. Details of the method are given in the next subsection.

3.2. Generative Language Model for Character Utterances

Our method consists of 2 steps: (1) cluster generation using character utterances, and (2) instruction-tuning dataset preparation using prompt engineering. The following subsections describe these steps in detail.

3.2.1. Character Clustering Using Utterances

We adopted the Project Dialogism Novel Corpus (PDNC) dataset [33], which consists of 28 novels, to construct a language model. In the PDNC dataset, speakers (i.e., characters) and utterances are provided in a structured form, making it easy to use for our purpose. We have converted the PDNC dataset into a set of character documents. Assume that the original PDNC dataset can be represented as a set D o = { ( c 1 , u 1 ) , ( c 2 , u 2 ) , }, where c i and u i represent the i-th character name and an utterance, respectively. Let C be a set of unique characters; note that | D o | is not necessarily equal to | C | . Given the original PDNC set, we created a set of character documents D c = { d 1 , d 2 ,…}, where d j denotes a set of all utterances of the j-th character in the character set C. Since this creates a document for each unique character, | D c | will be equal to | C | .
s ( i ) = b ( i ) a ( i ) m a x ( a ( i ) , b ( i ) )
Before clustering the character documents, the number of clusters is determined using Silhouette scores and the Elbow method. The character documents are represented in a document-word matrix, and the silhouette score s(·) is calculated using the Equation (1), where i is a data index, a(·) is a mean distance to the other data instances within the same cluster, b(·) is a mean distance to the other clusters. We got the optimal number of clusters is 2 using the silhouette score. We also used the elbow method, which finds the number of clusters using the sum of square errors (SSE), and found that the optimal number of clusters is 2 as shown in Figure 4. The K-means clustering algorithm with Euclidean distance is used to cluster the character documents into two groups. We denote the two clusters as M 1 and M 2 , and per-cluster character documents D M i = { d j d j M i }.
For each character cluster, the characters at the center have the strongest personality traits of the cluster, while the characters at the outer edges have the weaker personality traits. Using only the strongest characters at the center is not preferable because it suffers from the “unconscious author style” problem. However, using all characters within a cluster is also not a good choice because the outer characters are weak in representing the related personality. Therefore, we selected the top-k center characters for each cluster using silhouette scores, and merged or concatenated the top-k character documents into a single document. We define d M i k as the resulting document from the top-k character documents in the cluster M i .

3.2.2. Instruction-Tuning Dataset Preparation

We use the instruction-tuning approach to tune a decoder-based language model for translating character utterances according to a given personality. In the instruction-tuning approach a dataset is fed to the language model, so that the model generates the desired responses by following the instruction. An example of our instruction-tuning data is shown in Table 3. The instruction-tuning data consists of four parts: instruction, input, representative characters, and response. The instruction provides a general role or rules for the model to follow. The input and response are a neutralized utterance and an original utterance. The original utterance comes from d M i k , so this utterance has the personality of the cluster M i . On the other hand, the neutralized utterance is supposed to have no personality, but the same content as the original utterance. The model will learn how to translate the neutralized utterance (input) into the original utterance (response).
The representative characters are used to inform the model of a desired personality (or a target cluster). That is, instead of using a number or a meaningless symbol (e.g., cluster-01), we use some representative character names to refer to a target cluster.
There are two challenging issues in preparing the instruction-tuning dataset. First, we need to determine the representative characters of each cluster (i.e., the third part of the instruction-tuning data in Table 3). We assume that the representative characters only refer to a target cluster, so we simply take a single character with the best silhouette score.
Gabriel Syme (The Man Who Was Thursday) is chosen as the representative character for M 1 , and Mary Crawford (Mansfield Park) is for M 2 .
The second challenge is the neutralized utterance (i.e., the second part of the instruction-tuning data in Table 3). We design a few-shot prompt to ask a third-party generative language model, GPT-3.5-Turbo, to generate a neutralized utterance for a given utterance. Table 4 describes the prompt we design. Note that this prompt asks the third-party language model to work in exactly the opposite way that our model does; that is, the third-party language model converts original utterances (i.e., those with strong personality traits) into what we refer to as “neutralized” utterances. We use this term for convenience, but we do not claim that these outputs are truly neutral, only that they are comparatively less marked for personality traits than the originals. In the early stages of data transformation, no in-context examples were available; thus, we initially generated six utterances using a zero-shot prompting strategy. These initial outputs were then incorporated into the prompt as in-context examples, and the generation continued in a few-shot setting for the remaining data.
The preparation process for constructing the instruction-tuning dataset is as follows. As shown in Figure 5, an analysis of the word per sentence distribution revealed that sentences containing between 5~200 words accounted for the largest proportion, and we used data that fell within this range. We also chose this range because excessively short or long sentences were considered inappropriate for effectively conveying the character’s personality [38,39]. Figure 5 presents the following visualizations: the distribution of word counts for prompt sentences P t (top), the distribution of word counts for response sentences R s (bottom). Based on this, only the sentences with 5~200 words in d M i k were used, and the rest were filtered out. For each original sentence in d M i k , a neutralized utterance was generated using our designed prompt. We extracted representative character names of the character clusters, and the instruction-tuning dataset D i n s t was created according to the format in Table 3.

4. Experiment

4.1. Experiment Setup

Open-source decoder-based models are adopted and the model size (i.e., the number of trainable parameters) ranges from 7B to 13B. The models are Llama3.1-8B, Llama2-7B, Llama2-13B, Gemma2-9B, Falcon3-7B, and Falcon3-10B. We employed the AdamW optimizer [40], the batch size is 2, and the number of epochs for tuning is 3. Given the limited training data and computational resources, we adopted LoRA (Low-Rank Adaptation) [41] for efficient fine-tuning. The detailed hyperparameter settings used for training are provided in Table 5. We used the PyTorch library (version 2.2.2+cu121) and Hugging Face open-source models for implementation and training [42,43]. We accessed GPT-3.5-Turbo and GPT-4o via the OpenAI API to generate neutral utterances and to perform automatic evaluation, respectively. Our machine is equipped with RTX 3090 Ti (4EA) and 256 GB RAM. We arbitrarily set the number of center characters k (i.e., cluster size) to be 50, and the representative characters of the two clusters are Mary Crawford and Gabriel Syme. For inference, we set the temperatures 0.7 and 1.1 for Gabriel Syme and Mary Crawford, respectively. The dataset statistics used for instruction-tuning and model evaluation are detailed in Table 6. The training dataset consists of 144 utterances, and the test dataset used for evaluation contains 50 utterances.

4.2. Evaluation Methodology

To evaluate the effectiveness of our language model tuned by the instruction-tuning, we compare our tuned language model with its base model. Specifically, for each base model m b (e.g., Llama3.1-8b), we created its tuned model m t through the instruction-tuning. Each test prompt, as shown in Table 3 without the response part, is given to m b and m t , and their responses r m b and r m t are obtained. Recent studies have reported high evaluation consistency between human judgments and LLM-based evaluations [44,45], leading many LLM-related works to adopt Judge LLMs. In line with this trend, we used a strong LLM (e.g., GPT series) as the judge model in our evaluation. Our judge model, GPT-4o, is given our designed prompt, and makes a decision about which response among r m b and r m t is better in terms of the target personality.
Table 7 shows an example of the pair of r m b and r m t for the base model Llama3.1-8b.
We tested whether the judge model knew the personality of the target or representative characters (e.g., Mary Crawford and Gabriel Syme) by simply asking the judge model in natural language (e.g., Do you know the personality and characteristics “The Policeman” from The Man Who Was Thursday?), and found that the judge model knows their personality. Table 8 shows the judge model’s response.
We also randomly shuffled the order of r m b and r m t to deal with the position bias problem [46,47]. Another known issue in LLM-based evaluation is self-preference, where models tend to prefer responses they themselves have generated [48]. However, since this study uses a diverse set of models, it can be considered relatively free from this bias.

5. Results and Discussion

Table 9 summarizes the experimental results on four foundation models, averaged over three repeated trials. We observed that our tuned models, m t , were generally selected more often than the base models m b , suggesting that our models generated character utterances that better reflected the target character’s personality. There are two interesting observations here. First, our tuned Llama2-7B models showed the best performance, while the tuned Falcon3 models performed the worst, likely due to their primary training data source (i.e., RefinedWeb) [49] which is known to not contain full book or novel texts—potentially hindering their ability in literary domains. Second, the Mary Crawford cluster showed a relatively larger performance gap (i.e., Avg. #Win of m t vs. m b ) than the Gabriel Syme cluster. Table 10 shows this tendency more clearly. The reason may be that the Mary Crawford cluster has the personality that stands out more clearly; as shown in Figure 1, we see more noticeable changes in the translated utterances of Mary Crawford compared to the utterances of Gabriel Syme.
Our hypothesis is that our language models trained on the utterances of character ‘clusters,’ can translate a given utterance into a new utterance that better reflects the personality of a target character cluster.
Although we have shown that our language models outperform the base language models as shown in Table 9, we further investigated the question: “Does cluster size affect performance?” Specifically, we conducted experiments by varying the cluster size (i.e., the number of center characters k of the top-k character document d M i k ) from 10 to 73, and the result is shown in Table 11. We adopted Gemma2-9B as the base model. As shown in Table 11, we observed the best performance when k = 50 . This implies that using utterances with a small number of central characters (i.e., k = 10 ) is not preferable because it may cause the “unconscious author style” problem, and using utterances with all characters (i.e., k = 73 ) is also not a good choice because the outer characters are weak in representing the associated personality.

Limitations

A limitation of our method is that the tuned models sometimes produced utterances that were poor in terms of fluency. Figure 6 shows an example of the translation of a science fiction, Snow Crash [50]. Note that some of the translated utterances (e.g., “Oh, where do you be?”) are grammatically incorrect. The fluency problem is probably related to the size of the language models [51], and this limitation will be resolved as we tune larger models with larger datasets.
A further limitation is related to the clustering method used to represent character personality. While clustering based on personality trait annotations for each character would be ideal, it requires costly manual annotation. As a practical alternative, we adopted a word-based clustering approach. Notably, the resulting two clusters may reflect an implicit binary structure underlying character styles in the dataset. In future work, we plan to expand the dataset, explore a broader range of cluster numbers, and apply more sophisticated methods such as hierarchical clustering.

6. Conclusions

In this study, we aimed to address the problem of “unconscious author style” in order to build language models that generate character utterances more accurately reflecting a specific personality. To achieve this, we constructed the OCEAN-based character dataset in which trained literature experts annotated each character with scores on the five personality dimensions, an effort with significant academic value. In addition, we proposed a novel methodology and constructed a customized dataset that enables instruction-tuning using utterances from literary texts, providing a distinctive approach to character-centered language modeling. Through this approach, we also present a method for neutralizing the author’s unconsciously embedded stylistic tendencies while faithfully capturing the unique personality of each character. Based on this foundation, we fine-tuned several generative language models—including decoder-based models such as LLaMA 2, LLaMA 3, and Gemma 2—that have recently attracted considerable attention in the field of language modeling.
This study demonstrates the potential of using character utterances and instruction-tuned language models to effectively reflect the personality traits and individuality of fictional characters in literary texts. By combining novel-based utterance data with instruction-tuning for large language models (LLMs), we experimentally validate that character-specific expressions can be naturally generated in a manner consistent with their annotated personality traits. This work highlights the potential for interdisciplinary integration between natural language processing (NLP) and literary studies, and suggests meaningful directions in which language models can contribute to literary creativity and textual understanding. Moreover, this research serves as a practical example of how data-driven approaches yield tangible results in the convergence of the humanities and engineering. Future work will include extending the model to a wider range of novels to cover different writing styles and character profiles, as well as improving both quantitative and qualitative evaluation methods (e.g., stylistic consistency, literary realism, and etc.) for the generated text. Through such efforts, we expect to make meaningful contributions to AI-driven literary creation and the development of character-centered dialogue systems.

Author Contributions

Conceptualization, E.-J.K. and Y.-S.J.; methodology, Y.-S.J.; investigation, E.-J.K.; resources, M.Y. and C.-H.J.; writing—original draft preparation, E.-J.K.; writing—review and editing, E.-J.K., C.-H.J., M.Y. and Y.-S.J.; visualization, E.-J.K. and Y.-S.J.; supervision, Y.-S.J.; project administration, Y.-S.J.; funding acquisition, M.Y. and C.-H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Education of the Republic of Korea and the National Research Foundation of Korea (NRF-2022S1A5A2A03052880). This work was supported by the Hongik University new faculty research support fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

For inquiries regarding data sharing, please contact the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  2. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. Llama: Open and efficient foundation language models. arXiv 2023, arXiv:2302.13971. [Google Scholar]
  3. Gu, X.; Chen, M.; Lin, Y.; Hu, Y.; Zhang, H.; Wan, C.; Wei, Z.; Xu, Y.; Wang, J. On the effectiveness of large language models in domain-specific code generation. ACM Trans. Softw. Eng. Methodol. 2025, 34, 1–22. [Google Scholar] [CrossRef]
  4. Singhal, K.; Azizi, S.; Tu, T.; Mahdavi, S.S.; Wei, J.; Chung, H.W.; Scales, N.; Tanwani, A.; Cole-Lewis, H.; Pfohl, S.; et al. Large language models encode clinical knowledge. Nature 2023, 620, 172–180. [Google Scholar] [CrossRef] [PubMed]
  5. Singhal, K.; Tu, T.; Gottweis, J.; Sayres, R.; Wulczyn, E.; Amin, M.; Hou, L.; Clark, K.; Pfohl, S.R.; Cole-Lewis, H.; et al. Toward expert-level medical question answering with large language models. Nat. Med. 2025, 31, 943–950. [Google Scholar] [CrossRef] [PubMed]
  6. Garrido-Merchán, E.C.; Arroyo-Barrigüete, J.L.; Gozalo-Brizuela, R. Simulating HP Lovecraft horror literature with the ChatGPT large language model. arXiv 2023, arXiv:2305.03429. [Google Scholar]
  7. Rashkin, H.; Celikyilmaz, A.; Choi, Y.; Gao, J. Plotmachines: Outline-conditioned generation with dynamic plot state tracking. arXiv 2020, arXiv:2004.14967. [Google Scholar]
  8. Liu, D.; Li, J.; Yu, M.H.; Huang, Z.; Liu, G.; Zhao, D.; Yan, R. A character-centric neural model for automated story generation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1725–1732. [Google Scholar]
  9. Zhu, J.; Jurgens, D. Idiosyncratic but not arbitrary: Learning idiolects in online registers reveals distinctive yet consistent individual styles. arXiv 2021, arXiv:2109.03158. [Google Scholar]
  10. Jafariakinabad, F.; Hua, K.A. Style-Aware Neural Model with Application in Authorship Attribution. In Proceedings of the 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; pp. 325–328. [Google Scholar] [CrossRef]
  11. Leech, G.N.; Short, M. Style in Fiction: A linguistic Introduction to English Fictional Prose; Pearson Education: Harlow, UK, 2007; Number 13. [Google Scholar]
  12. Eisen, M.; Ribeiro, A.; Segarra, S.; Egan, G. Stylometric analysis of Early Modern period English plays. Digit. Scholarsh. Humanit. 2018, 33, 500–528. [Google Scholar] [CrossRef]
  13. Maslej, M.M.; Oatley, K.; Mar, R.A. Creating fictional characters: The role of experience, personality, and social processes. Psychol. Aesthet. Creat. Arts 2017, 11, 487. [Google Scholar] [CrossRef]
  14. Wu, W.; Wu, H.; Jiang, L.; Liu, X.; Hong, J.; Zhao, H.; Zhang, M. From role-play to drama-interaction: An LLM solution. arXiv 2024, arXiv:2405.14231. [Google Scholar]
  15. Zhang, S.; Dinan, E.; Urbanek, J.; Szlam, A.; Kiela, D.; Weston, J. Personalizing dialogue agents: I have a dog, do you have pets too? arXiv 2018, arXiv:1801.07243. [Google Scholar]
  16. Pennebaker, J.W.; King, L.A. Linguistic styles: Language use as an individual difference. J. Personal. Soc. Psychol. 1999, 77, 1296. [Google Scholar] [CrossRef] [PubMed]
  17. McCrae, R.R.; Costa Jr, P.T. Personality trait structure as a human universal. Am. Psychol. 1997, 52, 509. [Google Scholar] [CrossRef] [PubMed]
  18. Doyle, A.C. The Adventure of the Blue Carbuncle. In The Adventures of Sherlock Holmes; George Newnes: London, UK, 1892; Available online: https://www.gutenberg.org/ebooks/1661 (accessed on 29 April 2025).
  19. Doyle, A.C. A Scandal in Bohemia. In The Adventures of Sherlock Holmes; George Newnes: London, UK, 1892; Available online: https://www.gutenberg.org/ebooks/1661 (accessed on 29 April 2025).
  20. McCrae, R.R.; Costa, P.T. Validation of the five-factor model of personality across instruments and observers. J. Personal. Soc. Psychol. 1987, 52, 81. [Google Scholar] [CrossRef] [PubMed]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  22. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving language understanding by generative pre-training. 2018. Available online: https://openai.com/index/language-unsupervised/ (accessed on 29 April 2025).
  23. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  24. Team, G.; Anil, R.; Borgeaud, S.; Alayrac, J.B.; Yu, J.; Soricut, R.; Schalkwyk, J.; Dai, A.M.; Hauth, A.; Millican, K.; et al. Gemini: A family of highly capable multimodal models. arXiv 2023, arXiv:2312.11805. [Google Scholar]
  25. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar]
  26. Almazrouei, E.; Alobeidli, H.; Alshamsi, A.; Cappelli, A.; Cojocaru, R.; Debbah, M.; Goffinet, É.; Hesslow, D.; Launay, J.; Malartic, Q.; et al. The falcon series of open language models. arXiv 2023, arXiv:2311.16867. [Google Scholar]
  27. Team, G.; Mesnard, T.; Hardin, C.; Dadashi, R.; Bhupatiraju, S.; Pathak, S.; Sifre, L.; Rivière, M.; Kale, M.S.; Love, J.; et al. Gemma: Open models based on gemini research and technology. arXiv 2024, arXiv:2403.08295. [Google Scholar]
  28. Team, G.; Riviere, M.; Pathak, S.; Sessa, P.G.; Hardin, C.; Bhupatiraju, S.; Hussenot, L.; Mesnard, T.; Shahriari, B.; Ramé, A.; et al. Gemma 2: Improving open language models at a practical size. arXiv 2024, arXiv:2408.00118. [Google Scholar]
  29. Biderman, S.; Schoelkopf, H.; Anthony, Q.G.; Bradley, H.; O’Brien, K.; Hallahan, E.; Khan, M.A.; Purohit, S.; Prashanth, U.S.; Raff, E.; et al. Pythia: A suite for analyzing large language models across training and scaling. In Proceedings of the International Conference on Machine Learning. PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 2397–2430. [Google Scholar]
  30. Jiang, A.Q.; Sablayrolles, A.; Mensch, A.; Bamford, C.; Chaplot, D.S.; de Las Casas, D.; Bressand, F.; Lengyel, G.; Lample, G.; Saulnier, L.; et al. Mistral 7B. arXiv 2023, arXiv:2310.06825. [Google Scholar]
  31. Chowdhery, A.; Narang, S.; Devlin, J.; Bosma, M.; Mishra, G.; Roberts, A.; Barham, P.; Chung, H.W.; Sutton, C.; Gehrmann, S.; et al. Palm: Scaling language modeling with pathways. J. Mach. Learn. Res. 2023, 24, 1–113. [Google Scholar]
  32. John, O.P.; Robins, R.W.; Pervin, L.A. Handbook of Personality: Theory and Research, 3rd ed.; Guilford Press: New York, NY, USA, 2010. [Google Scholar]
  33. Vishnubhotla, K.; Hammond, A.; Hirst, G. The project dialogism novel corpus: A dataset for quotation attribution in literary texts. arXiv 2022, arXiv:2204.05836. [Google Scholar]
  34. Masumura, R.; Orihashi, S.; Ihori, M.; Tanaka, T.; Makishima, N.; Suzuki, S.; Mizuno, S.; Hojo, N. Multimodal Fine-Grained Apparent Personality Trait Recognition: Joint Modeling of Big Five and Questionnaire Item-level Scores. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 1456–1464. [Google Scholar]
  35. Cohen, J. Statistical Power Analysis for the Behavioral Sciences; Routledge: Oxfordshire, UK, 2013. [Google Scholar]
  36. Shrout, P.E.; Fleiss, J.L. Intraclass correlations: Uses in assessing rater reliability. Psychol. Bull. 1979, 86, 420. [Google Scholar] [CrossRef] [PubMed]
  37. Cicchetti, D.V. Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychol. Assess. 1994, 6, 284. [Google Scholar] [CrossRef]
  38. Nekvinda, T.; Dušek, O. AARGH! end-to-end retrieval-generation for task-oriented dialog. arXiv 2022, arXiv:2209.03632. [Google Scholar]
  39. Passali, T.; Chatzikyriakidis, E.; Andreadis, S.; Stavropoulos, T.G.; Matonaki, A.; Fachantidis, A.; Tsoumakas, G. From lengthy to lucid: A systematic literature review on NLP techniques for taming long sentences. arXiv 2023, arXiv:2312.05172. [Google Scholar]
  40. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  41. Hu, E.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. arXiv 2022, arXiv:2106.09685. [Google Scholar]
  42. Paszke, A. Pytorch: An imperative style, high-performance deep learning library. arXiv 2019, arXiv:1912.01703. [Google Scholar]
  43. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; pp. 38–45. [Google Scholar]
  44. Dubois, Y.; Galambosi, B.; Liang, P.; Hashimoto, T.B. Length-controlled alpacaeval: A simple way to debias automatic evaluators. arXiv 2024, arXiv:2404.04475. [Google Scholar]
  45. Peng, B.; Li, C.; He, P.; Galley, M.; Gao, J. Instruction tuning with gpt-4. arXiv 2023, arXiv:2304.03277. [Google Scholar]
  46. Liusie, A.; Manakul, P.; Gales, M.J.F. LLM Comparative Assessment: Zero-shot NLG Evaluation through Pairwise Comparisons using Large Language Models. arXiv 2024, arXiv:2307.07889. [Google Scholar]
  47. Wang, P.; Li, L.; Chen, L.; Cai, Z.; Zhu, D.; Lin, B.; Cao, Y.; Liu, Q.; Liu, T.; Sui, Z. Large language models are not fair evaluators. arXiv 2023, arXiv:2305.17926. [Google Scholar]
  48. Panickssery, A.; Bowman, S.; Feng, S. Llm evaluators recognize and favor their own generations. Adv. Neural Inf. Process. Syst. 2024, 37, 68772–68802. [Google Scholar]
  49. Penedo, G.; Malartic, Q.; Hesslow, D.; Cojocaru, R.; Cappelli, A.; Alobeidli, H.; Pannier, B.; Almazrouei, E.; Launay, J. The Refined-Web dataset for Falcon LLM: Outperforming curated corpora with web data, and web data only. arXiv 2023, arXiv:2306.01116. [Google Scholar]
  50. Stephenson, N. Snow Crash; Penguin Books: London, UK, 1992; p. 73. [Google Scholar]
  51. Kaplan, J.; McCandlish, S.; Henighan, T.; Brown, T.B.; Chess, B.; Child, R.; Gray, S.; Radford, A.; Wu, J.; Amodei, D. Scaling laws for neural language models. arXiv 2020, arXiv:2001.08361. [Google Scholar]
Figure 1. An example of utterance translation. The conversation between Holmes and Watson adapted from the Sherlock Holmes stories [18,19] (left) is translated into a conversation between two different characters (right). The lower table describes the personalities of the four characters by their per-trait keywords (all keywords can be found in Table 1).
Figure 1. An example of utterance translation. The conversation between Holmes and Watson adapted from the Sherlock Holmes stories [18,19] (left) is translated into a conversation between two different characters (right). The lower table describes the personalities of the four characters by their per-trait keywords (all keywords can be found in Table 1).
Applsci 15 08136 g001
Figure 2. Per-trait average OCEAN score distributions in our annotated dataset.
Figure 2. Per-trait average OCEAN score distributions in our annotated dataset.
Applsci 15 08136 g002
Figure 3. Graphical representation of two different clusters, where the dots are characters, the colors indicate personality-based clusters, and the position of the dots represent linguistic patterns of the characters.
Figure 3. Graphical representation of two different clusters, where the dots are characters, the colors indicate personality-based clusters, and the position of the dots represent linguistic patterns of the characters.
Applsci 15 08136 g003
Figure 4. The optimal number of clusters using the Elbow method, where the x-axis is the number of clusters, and the y-axis is the Within-Cluster Sum of Squares (WCSS).
Figure 4. The optimal number of clusters using the Elbow method, where the x-axis is the number of clusters, and the y-axis is the Within-Cluster Sum of Squares (WCSS).
Applsci 15 08136 g004
Figure 5. The distribution of words per sentence, where the x-axis represents the word count per sentence and the y-axis indicates the number of sentences corresponding to each word count. Specifically: word count distribution of prompt sentences P t (top) word count distribution of response sentences R s (bottom).
Figure 5. The distribution of words per sentence, where the x-axis represents the word count per sentence and the y-axis indicates the number of sentences corresponding to each word count. Specifically: word count distribution of prompt sentences P t (top) word count distribution of response sentences R s (bottom).
Applsci 15 08136 g005
Figure 6. An example of utterance translation. The conversation between Hiro Protagonist and Y.T in the novel Snow Crash [50] (left) is translated into a conversation between two different characters (right). The bottom table describes the personality of the four characters by their per-trait keywords (all keywords can be found in Table 1).
Figure 6. An example of utterance translation. The conversation between Hiro Protagonist and Y.T in the novel Snow Crash [50] (left) is translated into a conversation between two different characters (right). The bottom table describes the personality of the four characters by their per-trait keywords (all keywords can be found in Table 1).
Applsci 15 08136 g006
Table 1. Keywords representing low and high OCEAN traits scores [20].
Table 1. Keywords representing low and high OCEAN traits scores [20].
TraitLow Score Keyword ↔ High Score Keyword
Opennessconventional ↔ original
down-to-earth ↔ imaginative
uncreative ↔ creative
uncurious ↔ curious
prefers routine ↔ prefers variety
Conscientiousnesscareless ↔ careful
lazy ↔ hardworking
disorganized ↔ well-organized
undependable ↔ reliable
sloppy ↔ neat
Extraversionreserved ↔ talkative
aloof ↔ friendly
quiet ↔ sociable
retiring ↔ outgoing
passive ↔ active
Agreeablenessselfish ↔ selfless
rude ↔ courteous
irritable ↔ good-natured
suspicious ↔ trusting
stubborn ↔ flexible
Neuroticismcalm ↔ worrying
relaxed ↔ high-strung
secure ↔ insecure
comfortable ↔ self-conscious
hardy ↔ vulnerable
Table 2. Samples of OCEAN scores on novel characters.
Table 2. Samples of OCEAN scores on novel characters.
TitleCharacter NameOCEAN
Northanger AbbeyEleanor Tilney12−111
The Sport of the GodsJoe Hamilton2−12−12
Mansfield ParkHenry Crawford−2−232−1
Sense and SensibilityLucy Steele1−1211
A Passage to IndiaMr. McBryde−11−1−1−1
Table 3. Instruction-tuning data template.
Table 3. Instruction-tuning data template.
Format of Instruction Data
Instruction:
Transform the following neutral statement into the utterance of the Representative character.
Your response should reflect the Representative character’s unique tone, rhetorical style, and personality.
Do not repeat the input. Just write a single stylistically transformed response.
Input:
{Neutralized utterance}
Representative characters:
{Representative character of the book}
Response:
{Original utterance}
Table 4. A prompt designed to generate the neutralized utterances.
Table 4. A prompt designed to generate the neutralized utterances.
CategoryContent
PromptYou are a model that transforms input sentences by neutralizing personality in the original sentence.
Message (Excerpt)“You are generating data which will be used to train a machine learning model.”
“You will be given a {prompt} of the model we want to train, which is used to generate several response data samples based on each row of {sentence}.”
“You will do so in this format:”
FormatPrompt
______
{sentence}
______
Response
______
$response_goes_here
______
In the message“You should only generate one response per turn based on {sentence}.”
“For each turn, make the example slightly more complex than the last, while ensuring diversity.”
“Make sure your examples are unique and diverse, yet high-quality and complex enough to train a well-performing model.”
“Here is the type of model we want to train:”
Example (few-shot)Prompt
______
I can’t believe you did this! You’re always messing things up.
______
Response
______ It’s surprising that this happened. It seems like there’s a recurring issue.
______
Table 5. Hyperparameter settings used for instruction tuning with LoRA.
Table 5. Hyperparameter settings used for instruction tuning with LoRA.
HyperparameterValue
Learning rate5 × 10 5
Batch size (per device)2
Epochs3
OptimizerAdamW
LR schedulerlinear
LoRA target modulesq_proj, v_proj
LoRA rank4
LoRA alpha16
LoRA dropout0.05
Table 6. Statistics for word counts and standard deviations (Std.) of Prompts ( P t ) and Responses ( R s ) in the training and test datasets.
Table 6. Statistics for word counts and standard deviations (Std.) of Prompts ( P t ) and Responses ( R s ) in the training and test datasets.
# Chars/Cluster# UtterancesAvg. Words ( P t )Avg. Words ( R s )Std. Words ( P t )Std. Words ( R s )
Train dataset5014415.115.521.712.2
Test dataset-5022.822.126.817.7
Table 7. Example prompts and paired responses ( r m t and r m b ) generated by the instruction-tuned model m t and the base model m b , both based on Llama3.1-8b, to be judged by the GPT-4o model.
Table 7. Example prompts and paired responses ( r m t and r m b ) generated by the instruction-tuned model m t and the base model m b , both based on Llama3.1-8b, to be judged by the GPT-4o model.
No.ModelClusterPromptGenerated Response r m t Generated Response r m b
1Llama3.1-8bGabriel SymeI serve as the ultimate authority in detection matters. When Gregson, Lestrade, or Athelney Jones encounter challenges beyond their expertise, they bring the issue to me. I carefully analyze the information, providing a specialist’s perspective …As an expert detective, it is my duty to provide guidance when necessary. While I am content being behind the scenes, working tirelessly to uncover the truth, I take great pride in using my specialized knowledge and skills to help others succeed in their investigations.“You speak of recognition? It eludes me like an elusive thief.”
2Llama3.1-8bGabriel SymeNaturally, that would be the case.Certainly! That is exactly what I mean to say.Of course! It only stands to reason.
3Llama3.1-8bGabriel SymeIf he doesn’t come to me, then I’ll have to find another way to reach him.The police detective stood on his feet as if he were about to charge at some invisible enemy.It’s time for a plan B, my friend. If he refuses to play ball, we’ll just have to take matters into our own hands. We can’t let this case slip through our fingers; it’s too important! So gather up your wits and get ready for some detective work - because if he won’t help us, we’ll make sure that he does!
4Llama3.1-8bMary CrawfordIt’s imperative that we apprehend her.“My dear Emma,” Mary Crawford would say with a mischievous glint … “We must proceed swiftly lest anyone discovers our clandestine mission.”What is so important about apprehending her? If she doesn’t want to be apprehended, then why would anyone bother with it?
5Llama3.1-8bMary CrawfordI serve as the ultimate authority in detection matters. When Gregson, Lestrade, or Athelney Jones encounter challenges beyond their expertise, they bring the issue to me. I carefully analyze the information, providing a specialist’s perspective …Ah yes, dear Mary, serving as the master detective certainly brings its share of excitement! …People come knocking on my door all the time for help when things get tough …
6Llama3.1-8bMary CrawfordLady Catherine de Bourgh, is known for her strong personality and high expectations.“Ah! The idea of ‘expectation’ does cause some to lose sight of true values,” Mary commented with a mischievous glint …“A person who doesn’t respect themselves won’t receive it from others.”
Table 8. Verbatim answer regarding the character awareness of the judge model (GPT-4o).
Table 8. Verbatim answer regarding the character awareness of the judge model (GPT-4o).
TitleCharacterAuthorRoleCore TraitsSymbolism
The Man Who Was ThursdayGabriel SymeG.K. ChestertonUndercover detective seeking moral order within an anarchist council.Stoic, rational, insightful, morally firm, philosophicalRepresents hidden order beneath chaos.
Mansfield ParkMary CrawfordJane AustenCharming socialite whose wit and worldliness contrast the protagonist’s virtue.Witty, charming, worldly, independent, morally flexibleSymbol of modern secular appeal and moral ambiguity.
Table 9. Experimental results for m t versus m b .
Table 9. Experimental results for m t versus m b .
ModelSizeClusterAvg. #Win of m t Avg. #Win of m b
Llama27BGabriel Syme29.67 ± 1.1520.33 ± 1.15
7BMary Crawford36.67 ± 2.3113.33 ± 2.31
13BGabriel Syme22.67 ± 0.5827.33 ± 0.58
13BMary Crawford29.00 ± 1.7321.00 ± 1.73
Llama3.18BGabriel Syme27.00 ± 0.0023.00 ± 0.00
8BMary Crawford28.33 ± 1.5321.67 ± 1.53
Gemma29BGabriel Syme26.67 ± 0.5823.33 ± 0.58
9BMary Crawford28.67 ± 1.1521.33 ± 1.15
Falcon37BGabriel Syme19.67 ± 0.5830.33 ± 0.58
7BMary Crawford21.33 ± 0.5828.67 ± 0.58
10BGabriel Syme19.00 ± 2.0031.00 ± 2.00
10BMary Crawford20.00 ± 1.0030.00 ± 1.00
Table 10. Total win counts and win rates (%) of m t and m b for each cluster.
Table 10. Total win counts and win rates (%) of m t and m b for each cluster.
ClusterSum #Win of m t Sum #Win of m b Win Rate of m t Win Rate of m b
Gabriel Syme43446648.22%51.78%
Mary Crawford49240854.67%45.33%
Table 11. Evaluation results of Gemma2-9B, showing win counts for the tuned model ( m t ) and base model ( m b ) across different TopK values.
Table 11. Evaluation results of Gemma2-9B, showing win counts for the tuned model ( m t ) and base model ( m b ) across different TopK values.
TopK#Win of m t #Win of m b
TotalGabriel SymeMary CrawfordTotalGabriel SymeMary Crawford
10492029513021
30522428482622
50532429472621
73472522532528
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, E.-J.; Joe, C.-H.; Yun, M.; Jeong, Y.-S. Generative Language Models for Personality-Based Utterances in Novels: A Character Clustering Approach. Appl. Sci. 2025, 15, 8136. https://doi.org/10.3390/app15158136

AMA Style

Kim E-J, Joe C-H, Yun M, Jeong Y-S. Generative Language Models for Personality-Based Utterances in Novels: A Character Clustering Approach. Applied Sciences. 2025; 15(15):8136. https://doi.org/10.3390/app15158136

Chicago/Turabian Style

Kim, Eun-Jin, Chung-Hwan Joe, Misun Yun, and Young-Seob Jeong. 2025. "Generative Language Models for Personality-Based Utterances in Novels: A Character Clustering Approach" Applied Sciences 15, no. 15: 8136. https://doi.org/10.3390/app15158136

APA Style

Kim, E.-J., Joe, C.-H., Yun, M., & Jeong, Y.-S. (2025). Generative Language Models for Personality-Based Utterances in Novels: A Character Clustering Approach. Applied Sciences, 15(15), 8136. https://doi.org/10.3390/app15158136

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop