Next Article in Journal
General Neutral Functional Differential Equations of Third Order: Enhanced Oscillation Criteria
Previous Article in Journal
Handheld 3D Scanning-Based Robotic Trajectory Planning for Multi-Layer Multi-Pass Welding of a Large Intersecting Line Workpiece with Asymmetric Profiles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Cross-Prompt Essay Trait Scoring via External Knowledge Similarity Transfer

1
School of Computer and Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
2
School of Information Engineering, Minzu University of China, Beijing 100081, China
3
National Language Resource Monitoring and Research Center of Minority Language, Minzu University of China, Beijing 100081, China
4
Chinese Character Research and Application Laboratory, Beijing Normal University, Beijing 100875, China
5
Linguistic Data Consortium, University of Pennsylvania, Philadelphia, PA 19104, USA
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(5), 739; https://doi.org/10.3390/sym17050739
Submission received: 18 April 2025 / Revised: 1 May 2025 / Accepted: 9 May 2025 / Published: 11 May 2025
(This article belongs to the Section Computer)

Abstract

:
Cross-prompt automated essay scoring presents a significant challenge due to substantial differences in samples across prompts, and recent research has concentrated on evaluating distinct essay traits beyond the overall score. The primary approaches aim to enhance the effectiveness of AES in cross-prompt scenarios by improving shared representation or facilitating the transfer of common knowledge between source and target prompts. However, the existing studies only concentrate on the transfer of shared features within essay representation, neglecting the importance of external knowledge, and measuring the degree of commonality across samples remains challenging. Indeed, higher similarity of external knowledge also results in a better shared representation of the essay. Based on this motivation, in this paper, we introduce an extra-essay knowledge similarity transfer to assess sample commonality. Additionally, there is insufficient focus on the intrinsic meaning of the traits being evaluated and their varied impact on the model. Therefore, we incorporate extra-essay knowledge representation to enhance understanding of the essay under evaluation and the target of the task. Experimental results demonstrate that our approach outperforms other baseline models on the ASAP++ datasets, confirming the effectiveness of our method.

1. Introduction

Automated essay scoring (AES) aims to automatically assess an essay based on a provided prompt that presents the writer with the topic and often includes specific guidelines or instructions. The AES system has the potential to generate objective and equitable results swiftly and efficiently, addressing the issue of resource consumption and strong subjectivity in manual evaluation [1,2,3]. Initial studies in AES focused on providing a holistic evaluation by assigning an overall score to the essay [4,5,6,7,8] and gradually shifted to trait scoring from different perspectives [9,10] as the functionality expanded and achieved the objective of delivering multidimensional analysis to improve reliability. Currently, the research focus has changed from prompt-specific [6] to more challenging cross-prompt scenarios [11,12,13,14,15,16], as depicted in Figure 1.
Cross-prompt AES involves the evaluation of previously unseen target prompts using a training model from the source prompt. In this paper, we focus on the cross-prompt essay trait scoring (CPETS) where both the overall score and detailed trait scores in the target domain are provided, as illustrated in Figure 1b. Existing CPETS methods improve effectiveness by improving shared representation or facilitating the transfer of common knowledge between source and target prompts through content-related features [17,18] or external knowledge [1,19]. Ridley et al. [17] proposed a model for CPETS, which utilizes part-of-speech (POS) embeddings as input for a neural AES model and concatenates features such as readability over essay representation. Sun et al. [18] fused syntactic embedding with POS embedding to extend the essay representation, leveraging common linguistic knowledge to enhance the essay representation of the target domain. Do et al. [19] enhanced essay representation through incorporating prompt information and alleviated data insufficiency through extracted topic coherence features. They also endeavored to improve the model by quantifying relations between traits. While the studies above improved essay-shared representation by integrating various features, they devoted less attention to knowledge beyond the essay content, such as genre features and the intrinsic meanings of traits. Moreover, they did not focus on quantifying knowledge transfer across prompts. Knowledge transfer provides methodological support for cross-domain research, and knowledge flows across domains through techniques such as transfer learning and domain adaptation [20]. Chen and Li [1] concentrated on the challenge of knowledge transfer across distinct prompts, arguing that similar prompts can gain more shared knowledge.
As shown in Table 1, Prompt 1 and Prompt 2 were both argumentative essays written to require a discussion of a situation related to the academics of the students, and Prompt 3 was a narrative essay. As can be seen in the essays for Prompt 1 and Prompt 2, both essays are organized as question-and-answer writing with a high degree of vocabulary similarity. However, there were fewer commonalities between Prompt 3 and Prompts 1 and 2 at the organizational level of the essays, as well as greater differences in vocabulary use and language expression. It can be seen that essays with similar prompts or similar genres are more likely to contain similar information, even though the different essays come from different writers. Overall, the rational use of external information can enhance the ability of the model to extract shared features between source and target prompts and significantly improve the effectiveness of the model in scoring cross-prompt essays, further illustrating the effectiveness of solving the problem from the perspective of external information. Despite the study’s success, its focus was limited to the dissimilarities between prompts. The researchers did not explore the transferability of other dimensions of external knowledge between the source and target prompts. In addition, the effective measurement of the degree of commonality between knowledge and the efficient transfer of knowledge are crucial.
To resolve the aforementioned problems, we present cross-prompt trait scoring based on extra-essay knowledge transfer (ExtCTS). For the first problem, inadequate attention to knowledge beyond the content of the essay, we introduce extra-essay knowledge representation. This enhancement augments the understanding in the model of both the prompt under evaluation and the intrinsic meaning of the assessed trait by integrating the essay prompt and the trait description. Description refers to the explanation of a specific essay trait in a natural language text. For instance, a trait description may illustrate the evaluation of the trait “narrativity”, defined as “a measure of the coherence and cohesion of the response to the prompt”. Genre description, prompt description, and trait description details are provided in [21]. To tackle the problem of insufficient sources for knowledge transfer and the complexity of assessing commonality, we propose extra-essay similarity transfer, which involves evaluating the similarity of different external knowledge in source and target domains, setting similarity factors to determine transferability, and applying it to essay representation. The main contributions of our work are summarized as follows:
  • We propose extra-essay knowledge representation, which consists of an essay prompt encoder and a trait description encoder to integrate knowledge beyond the content of the essay.
  • We introduce extra-essay similarity transfer, which quantifies the similarity between various prompts by considering knowledge similarity factors beyond essay content, thereby enhancing essay representation.
  • Experimental results demonstrate that our approach achieved state-of-the-art performance, validating the effectiveness of the proposed methodology.
The paper is organized as follows. In Section 2, we first present the literature review. We describe our method and experiments in Section 3 and Section 4. Finally, in Section 5, we provide a conclusion, future work, and the limitations of the paper.

2. Related Work

In this section, we briefly review relevant studies concerning cross-prompt holistic scoring and cross-prompt essay trait scoring.

2.1. Cross-Prompt Holistic Scoring

Presently, AES is mostly performed as supervised learning, demanding substantial amounts of labeled data. However, the process of labeling data is time-consuming and labor-intensive, resulting in a shortage of training data for identical prompts. Phandi et al. [11] initiated research on cross-prompt holistic scoring to address the challenges. Phandi et al. [11] and Dong and Zhang [22] test the impacts of traditional AES and neural AES in cross-prompt settings on the ASAP [available online: https://www.kaggle.com/c/asap-aes (accessed on 1 May 2025)] dataset, respectively, establishing the foundational framework for subsequent research.
Subsequent research has focused on refining frameworks [23,24,25] and improving methods of cross-prompt knowledge transfer [26,27,28], utilizing multi-task learning [29], or transfer learning [30,31] in natural language processing. To tackle the challenge of extracting highly transferable scoring knowledge from multiple prompt sources, Li et al. [28] integrated prompt information and extracted diverse knowledge essential to the target domain through the fusion of data from all source prompts. Cao et al. [26] proposed a domain-adaptive framework to address the issue of neglecting domain differences in prior studies, optimize the utilization of data under various prompts, and enhance the domain-adaptive capabilities of models. Jiang et al. [27] developed disentangled representation learning to tackle the challenge of cross-domain generalization in essay representation. They further devised a prompt-aware neural AES model to extract the combined representation of prompt-invariant and prompt-specific features. These studies have enhanced the representation and transfer of knowledge across prompts. However, they have not explored knowledge transfer across traits.

2.2. Cross-Prompt Essay Trait Scoring

Cross-prompt essay trait scoring that evaluates not only the essay as a whole but also the individual trait was first proposed by Ridley et al. [12]. They proposed a CTS model that evaluates multiple attributes using a multi-task learning framework. The model utilizes POS embeddings as input and concatenates features like readability in essay representation. Furthermore, they improved representation by incorporating inter-trait relationships. Sun et al. [18] fused syntactic features and contrastive learning to enhance model performance by augmenting trait representation and improving model stability. Although both studies introduced linguistic knowledge, such as lexicon and syntax, at the input layer, they fell short in enhancing the utilization of external knowledge of essay across prompts. Do et al. [19] employed prompt information and topic coherence features to enhance essay comprehension of evaluation topics. They also established a loss function for trait relations by utilizing the correlation between different traits to improve the representation of the essay. However, they did not focus on quantifying knowledge transfer across prompts. Chen and Li [1] concentrated on the challenge of knowledge transfer across distinct prompts, arguing that the similarity in essay representations reflects the extent of transferable features. Consequently, they proposed that prompt-mapping captures the consistent representation of the target and source prompts. Zhang et al. [16] proposed a PDA framework, which effectively aligns the features of source and target prompts by using two alignment strategies, such as prompt-level alignment and category-level alignment, and further improves the performance of the model through classifier consistency constraint. Nevertheless, they only focus on the transfer of shared features within essay representation, neglecting the transfer of external knowledge. We believe that evaluating the degree of commonality of different dimensions of knowledge between source and target prompt samples and quantifying transferable knowledge are crucial directions for addressing cross-prompt trait essay scoring.

3. Methodology

In this paper, we focus on cross-prompt essay trait scoring. When given a dataset of n essay prompts, one is selected from the dataset as the target prompt P T and the remaining prompts are used as the source prompt set P E = P 1 , P 2 , , P i , , P n 1 ; the goal of cross-prompt traits scoring is to train a model using the source prompt P E and test it on the target prompt P T . The ith prompt P i contains N i examples of essays with annotations x i j , Y i j j = 1 N i , where x i j contains the text of the jth essay e i j and the external information of the jth essay K i j , Y i j = y i j 1 , y i j 2 , , y i j K is the set of scores to different traits for the jth essay, and K is the number of traits to be scored. The external information K i j consists of the prompt text p i j for the ith prompt, the genre text g i j for the ith prompt, the trait t r i j to be evaluated for the ith prompt, and its corresponding trait description t d i j , etc.
Existing cross-prompt essay trait scoring studies have focused on improving shared representations or facilitating the transfer of common knowledge between source and target prompts to improve the effectiveness of AES in cross-prompt scenarios. In this paper, we introduce the transfer of external essay knowledge similarity to assess sample commonality, which to some extent reflects symmetry considerations because when external knowledge has higher similarity, the shared representation of essays is also better, which implies that there is a potential symmetrical correlation between essays with different prompts in terms of external knowledge dimensions, reflecting the dynamic balancing mechanism in the process of knowledge transfer, and through which the grasp of essay commonality can be enhanced. Meanwhile, previous studies have paid insufficient attention to the intrinsic meanings of the assessed features and their varying impacts on the model. In contrast, this paper introduces additional essay knowledge representation, which enables the model to have a more in-depth understanding of the essays to be assessed and the task objectives. This approach also helps to achieve a more balanced and comprehension of the features across essays with different prompts, thereby establishing a kind of symmetry in feature understanding.
The proposed ExtCTS model is shown in Figure 2. Figure 3 is the sub-module of ExtCTS. It contains three components: extra-essay knowledge representation (EKR), multilevel extra-essay knowledge similarity transfer (EST), and a trait-specific encoder. EKR represents external knowledge to enhance the model to understand the prompt under evaluation and the trait being assessed. EST quantifies the degree of commonality among various external knowledge between the source and target domains. The factor from EST is then applied to the trait-specific encoder, enhancing the ability of the model to transfer knowledge from diverse prompts to obtain superior essay representation.

3.1. Extra-Essay Knowledge Representation

In order to enhance the ability of the model to extract shared features of the target prompts through external information, this section incorporates the trait description information under the framework of multi-trait scoring with fused prompt information. The following describes the EKR used in this paper to extract the external information representation, including the trait description encoder (TDE) and the essay prompt encoder (EPE).

3.1.1. Trait Description Encoder

Trait description is a piece of natural language that details and describes an evaluation trait. The TDE enhances the understanding of evaluation traits by fusing trait description information. For the traits t r 1 , t r 2 , , t r i , , t r K included in the training set, K is the number of traits, t r i is the ith target trait, and the trait descriptions are t d 1 , t d 2 , , t d i , , t d K . The ith trait description t d i has the word sequence t w i 1 , t w i 2 , , t w i n , n is the number of words, and obtains the corresponding word embedding vector as tx i 1 , tx i 2 , , tx i n . All trait descriptions are sequentially input to the word encoder (WE), and the model structure is shown in Figure 3a. The word encoder captures sentence-level representations of the input sequence using a CNN that incorporates an attention mechanism. Then the sentence-level trait description representation ts = ts 1 , ts 2 , , ts K is obtained by computing the Equations (1)–(4), and the formulas are shown below:
c i k = f W c · tx i k : tx i k + h w 1 + b c ,
z i k = tanh W z · c i k + b z ,
α i k = exp w α · z i k exp w α · z i j ,
ts i = α i k c i k ,
where W c and b c are the CNN trainable weight matrix and bias, respectively, h w is the window size of the CNN, W c is used to capture the features of all the possible consecutive h w words in the sentence, f · is the nonlinear activation function, and W z , w α , and b z are weight matrices, trainable weight vectors, and bias for the attention mechanism. The sentence-level trait description representation ts is computed from the contextual features c i of the CNN output. Attention vectors z i and attention weights α i are computed by Equation (2) and Equation (3), respectively.
Next, the inter-trait relations are further captured through the multi-head self-attention layer and LSTM with an attention mechanism, thus obtaining the final trait description representation. The sentence-level trait description representation ts is input to the multi-head self-attention layer and the formulas are shown below:
MSA ( q , k , v ) = H 1 H i H n W o ,
H i = Attention q W q i , k W k i , v W v i ,
where W q i , W k i , W v i are the parameter matrices of the ith head q = k = v = ts , Attention is the scaled dot product attention, ⊕ is concatenation, H i is the ith head used in the multi-head self-attention layer, n is the number of heads in the multi-head self-attention layer, and W o is the parameter matrix of the multi-head self-attention layer. Then, an LSTM is used to obtain the trait description representation d = d 1 , d 2 , , d K , which is computed as shown in Equation (7):
d u = LSTM o u 1 , o u ,
where o is the output of the multi-head self-attention layer and d u denotes the hidden state vector at the uth time.

3.1.2. Essay Prompt Encoder

In order to enhance the understanding of the model on the evaluation prompt and obtain more consistent evaluation results with the prompt information, this section uses a Word+POS Encoder that fuses word sequences and POS sequences to represent the essay prompt information, and the model structure is shown in Figure 3b. For the training set containing essay prompt texts p 1 , p 2 , , p i , , p J , J is the total number of prompts, the ith essay prompt text p i has the word sequences p w i 1 , p w i 2 , , p w i m , and obtains the corresponding word embeddings px i 1 , px i 2 , , px i m and the POS embeddings pp i 1 , pp i 2 , , pp i m . First, the word embeddings and POS embeddings are summarized px i 1 + pp i 1 , px i 2 + pp i 2 , , px i m + pp i m . Second, the summed embedding vectors are sequentially input to the CNN with attention mechanism to capture the sentence-level prompt features using the Equations (1)–(4). Then, the sentence-level prompt features are input to the multi-head self-attention layer and LSTM computed using Equations (5)–(7) to obtain the essay prompt representation h = h 1 , h 2 , , h J . Finally, the attention mechanism is computed for the essay prompt representation using Equations (8)–(10) to obtain the essay prompt contextual representation h :
z i = tanh W z · h i + b z ,
α i = exp w α · z i exp w α · z j ,
h = α i h i ,
where W z , w α and b z are the weight matrices, trainable weight vectors, and bias of the attention mechanism. The essay prompt contextual representation h is computed from the LSTM output of the essay prompt representation h . The attention vector z and the attention weight α are computed by Equation (8) and Equation (9), respectively.

3.2. Multi-Level Extra-Essay Knowledge Similarity Transfer

Essays with a high degree of external knowledge similarity are more likely to help the model learn better-quality shared features. Based on this motivation, we design multilevel similarity transfer to measure the degree of similarity between external information of the source prompts and the target prompt. First, the features of external knowledge are extracted using a pre-trained language model. Given a dataset of N p prompts, there is the set of external knowledge K = K g e n r e , K p r o m p t , K t r a i t , where K g e n r e = g i i = 1 N g is the set of N g genre descriptions, K p r o m p t = p i i = 1 N p is the set of N p essay prompts, K t r a i t = t d i i = 1 N t is the set of N t d trait descriptions. Given a source prompt training data e S with external knowledge g S , p S , and a target prompt with external knowledge g T , p T , the representation of the external knowledge of the essay is extracted using BERT with the following formula:
H = BERT [ C L S ] c [ S E P ] ,
where c is the input of BERT, ⊕ is concatenation operation. Then the cosine similarity of the external knowledge between the source prompt H S and the target prompt H T is calculated to obtain the similarity transfer factor; the formula is as follows:
cosine H S , H T = H S · H T H S H T .
In this paper, we consider two similarity factors, f s i m = f p , f g , where f p is the similarity factor for essay prompt and f g is the similarity factor for genre description. Finally, the features at different levels of trait-specific encoder are weighted using the f s i m as shown in Figure 2, so as to enable the source prompt features extracted by the neural network to transfer directionally towards the target prompt and to enhance the ability of the model to generalize to the target prompt.

3.3. Trait-Specific Encoder

We employ a hierarchical structure [4] and basic framework from [19]. The model uses POS encoder to obtain sentence-level representation as shown in Figure 3c and trait-specific encoder for each trait to obtain document-level representation as illustrated in Figure 2. For a sentence containing n words w 1 , w 2 , , w n , the POS embedding p 1 , p 2 , , p n is first obtained, and then an attention layer is used to obtain the sentence-level representation s . The formulas are as follows:
c i = CNN p i : p i + h w 1 , i = 1 , 2 , , n ,
s = Attention c 1 : c n ,
where h w is the kernel size of CNN.
Then, sentence-level essay representation is extracted using a text encoder as depicted in Figure 3d and further incorporate prompt representation. In addition, to enhance awareness of the target prompt, we calculate the document-level essay representation. The final representation of the model is t i , which is weighted by f s i m , and the score is obtained using the following equation:
y ^ i = σ W y i · f s i m · t i + b y i ,
where σ is the linear layer of sigmoid activation, W y i and b y i are the trainable weight matrices and bias terms, and y ^ i is the score predicted for the ith trait; t i contains the output of trait-specific encoder and all external knowledge representations obtained from TDE and EPE.

3.4. Loss Function

Loss function includes score loss L s c o r e and inter-trait similarity loss based on trait description L s i m , where L s c o r e uses the same strategy as [12].
L = η · L s c o r e + 1 η · L s i m ,
L s c o r e = 1 N 1 K i = 1 N j = 1 K y ^ i j y i j 2 ,
L s i m = 1 C i = 2 K j = i + 1 K TDS y ^ i , y ^ j , H t d ( i ) , H t d ( j ) ,
TDS = 1 cos y ^ i , y ^ j , if cos H t d ( i ) , H t d ( j ) δ 0 , otherwise ,
where η is a manual hyperparameter; N is the number of essays, and K is the number of traits; cos is the cosine similarity, δ is the similarity threshold, and C is the number of times the similarity of the trait description is greater than the threshold; y ^ i = y ^ 1 i , y ^ 2 i , , y ^ N i is the prediction of the ith trait of N training data; H t d ( i ) is the trait description for the ith trait.

4. Experiments

4.1. Datasets and Settings

We conducted experiments on the ASAP++ dataset [21], initially introduced by Ridley et al. [17] for CPETS, which was labeled with fine-grained trait scores on the ASAP [available online: https://www.kaggle.com/c/asap-aes (accessed on 1 May 2025)] dataset. The statistical results are presented in Table 2. Compared with overall scoring, trait scoring mainly evaluates the quality of essays in terms of language quality (vocabulary, syntax, etc.), content quality (strength of argument), and thinking level (organization). Evaluating the quality of essays from multiple dimensions can reflect writing level more comprehensively and promote the comprehensive improvement of the writing ability of students.
Additionally, other resources like genre descriptions, prompt descriptions, and trait descriptions derived from expert annotations are detailed in [21]. Prompt descriptions are shown in the Table 3. Genre descriptions are shown in the Table 4. For evaluation, we employed quadratic weighted kappa (QWK) [available online: https://www.kaggle.com/c/asap-aes#evaluation. (accessed on 1 May 2025)], a widely used metric in existing AES research [2]. QWK assesses the consistency between the AES system results and the golden score and assigns a value between 0 , 1 . QWK can be calculated by Equations (20) and (21):
W i , j = ( i j ) 2 ( N 1 ) 2 ,
QWK = 1 i , j W i , j O i , j i , j W i , j Z i , j ,
where O i , j denotes the number of essays that receive a rating i by the human rater and a rating j by the AES system. The expected rating matrix Z is histogram vectors of the golden rating and AES system rating and normalized so that the sum of its elements equals the sum of its elements in O ; N is the range of score for each essay.
We use the same data settings as Ridley et al. [12], who first proposed the CPETS task, with the data in the target domain coming from a prompt that does not appear in the training set and the other prompts serving as the source domain. We use the same data settings as Ridley et al. [12]. We employ the handcrafted features proposed by Do et al. [19], encompassing length-based, readability, text complexity, text variation, sentiment, and topic coherence features. We employ the handcrafted features proposed by Do et al. [19]. We utilize word 50-dimension embedding as input from [32]. We build POS 50-dimension embedding and train all models for 50 epochs with a batch size of 32. The output size of the CNN layer and the LSTM layer are both configured at 100. The window size for the CNN layer is set to 5, and we employ two heads for the multi-head self-attention. Each model is executed five times with the seeds { 12 , 22 , 32 , 42 , 52 } , following Ridley et al. [12], and the average QWK is reported. We utilize the RMSProp optimizer, setting the initial learning rate to 0.001 and the dropout rate to 0.5. δ is set to 0.7 and η is set to 0.7. Our experiments are conducted using an NVIDIA GeForce GTX 1080 Ti GPU.

4.2. Comparisons

We compare our method with the existing CPETS models:
  • Hi att [4]: This model was first applied to CPETS by Ridley et al. [12]. The model utilizes a hierarchical structure that uses a convolutional neural network with attention and a recurrent neural network with attention to model word-level and sentence-level representation.
  • AES aug [33]: The model builds on [6] by simultaneously evaluating multiple traits through multi-task learning.
  • PAES [17]: This model employs POS embedding as input, employing a hierarchical structure with distinct models dedicated to each trait.
  • CTS-w/o att [12]: This model removes the trait-level attention layer compared to CTS.
  • CTS [12]: This model is based on [17], which designs shared encoder modeling word-level representation, builds submodules for each trait, and enhances the representation by exploiting the relationships between different traits via trait-level attention.
  • FEATS [34]: This model is based on the handcrafted features produced by Ridley et al. [17] and Uto et al. [35], with the addition of POS bigram, prompt adherence features, top-N words features, and pronoun features.
  • PMAES [1]: This model improves representation by facilitating knowledge transfer through prompt-mapping, thus acquiring more shared knowledge from similar prompts.
  • RDCTS [18]: This model integrates POS embedding and syntactic embedding as inputs, leveraging contrastive learning to address data insufficiency.
  • PLAES [36]: This model captures shared knowledge between different source prompts through meta-learning and utilizes level-aware learning to improve the ability of model to distinguish between different levels of essay.
  • ProTACT [19]: This model utilizes prompt information and prompt coherence features to enhance the understanding of essays within the evaluation prompt.

4.3. Main Results

This section describes the overall effectiveness of the model proposed in this paper on a cross-prompt multi-trait task. The experimental results are summarized in terms of prompts and traits, and the average results for each prompt are shown in Table 5, while the results for each trait are shown in Table 6. The table contains two scenarios proposed in this paper, ExtCTS using BERT(base) and ExtCTS using BERT(large) to extract external information for calculating the similarity transfer factor. From Table 5 and Table 6, it can be seen that the ExtCTS model obtains the best results in terms of average QWK values for all prompts and traits. In particular, the ExtCTS model using BERT(base) to extract external information has average QWK values of 0.602 and 0.598 for all prompts and traits, respectively, which are 1% and 1.2% higher compared to the best baseline model. The ExtCTS model using BERT (large) to extract external information has average QWK values of 0.601 and 0.598 for all prompts and traits, respectively, which are 0.9% and 1.2% higher than the best baseline model.The above results show that ExtCTS is effective.
Comparing the results for each prompt, the two ExtCTS models were only slightly lower than the ProTACT model for Prompt 7, and the other results were superior to all baseline models. Compared with the best baseline model, ProTACT, the two ExtCTS gain about 2% improvement on prompts 1, 3, 4, and 6. Comparing the two different external information extraction schemes, ExtCTS using BERT(base) works better on Prompts 4, 5, 6, and 8, and ExtCTS using BERT(large) works better on Prompts 1 to 3. The above results show that the proposed method can enhance the effectiveness of each prompt set, and both external information feature extraction using BERT(base) and BERT(large) are effective.
Comparing the results for each trait, the best results for ExtCTS outperform the baselines on all traits, with only ExtCTS using BERT(large) being slightly lower than the best baseline on Organization. The two pre-trained models have similar results in the attribute dimension. Comparing ExtCTS with the best baseline, ExtCTS gains 2.5%, 2%, 1.6%, and 1.6% in Sentence Fluency, Word Choice, Content, and Prompt Adherence, respectively. The above results show that our method is able to obtain great enhancement in each trait. ExtCTS has more improvement in the traits of Sentence Fluency and Word Choice which are related to language expression, indicating that our method is able to better obtain the shared features from the source prompt that are related to language expression. ExtCTS has more enhancement in the trait of Content, and Prompt Adherence also gained good improvement, indicating the effectiveness of the fusion of external knowledge. However, the two traits of Organization and Conventions are not very sensitive to external knowledge. Organization focuses on internal text structure, such as paragraphing and levels of argumentation, which can be assessed directly through the analysis of contextual relationships in the model. Conventions relies on grammatical rules and standardized writing conventions and is less influenced by subject matter or domain knowledge.

4.4. Ablation Studies

In order to further validate the effectiveness of our method, ablation studies are conducted in this section, and the experimental results are shown in Table 7. The average QWK values for the ExtCTS model using BERT(base) over all prompts are given, the results with different external information removed are given in rows 1–5, and the results with different levels of transfer removed are given in rows 6–11. First, the effect of different external information on the model is verified. Comparing lines 1 to 4 of Table 7, removing h , d , and s , the model performances all decrease, with s decreasing the most. It shows that using external knowledge can effectively enhance the model performance. Comparing lines 2 to 5 of Table 7, the models using the above three representations are all able to enhance, further demonstrating the effectiveness of fusing external knowledge.
Then, the effect of different layers of similarity transfer on the model is verified. Comparison of Table 7, lines 1 and 6 to 8 shows that the results of removing the shallow level feature transfer have been improved, and removing the middle and deep level feature transfer have been decreased. The above results indicate that the factor used at the shallow level is less effective; most likely, the model mainly contains sentence-level features at the shallow level, which are more semantically different from the deep-level features that are directly used for scoring, and thus it is more difficult to transfer features from the shallow level. In addition, migrating at a shallow level of the model may further affect the training of the overall representation of the model, thus degrading the model performance. Comparison of rows 1 and 9 to 11 of Table 7 shows that the results of using only similarity feature transfer at the deeper level of the model decreased less, and the results of using only shallow transfer decreased more and were lower than the results of the best baseline model, further suggesting that transfer at a shallow level of the model affects the model results. In contrast, similarity feature transfer at the deeper level of the model is more effective, indicating that our method is more suitable for feature transfer at the deeper level of the model.

4.5. Effect of Loss Function

This section analyzes the impact of the similarity loss function between traits proposed in this paper (TDS), and the experimental results are shown in Table 8. When η takes the value of 1, the training objective of the model contains only the scoring loss, and it can be seen that the effect of the model using only the scoring loss is slightly degraded; the results are slightly better only in Prompt 5, which proves the validity of the inter-trait similarity loss function.
Next, the trait similarity (TS) loss function proposed by the baseline [19] is compared. The difference between the two lies in the calculation of the similarity threshold, and our method proposes to utilize the cosine similarity between different trait descriptions instead of the Pearson correlation coefficient. The experimental results are shown in Figure 4 and Figure 5. The TDS loss proposed in this paper is closer to the results of the baseline modeling method, which proves that the proposed method is competitive.

4.6. Effect of Transfer Factors

This section verifies the effect of the two trait similarity factors of f p and f g , and the results are shown in Table 9, which shows that the effect of using the essay prompt similarity factor is better. The genre similarity factor has a negative effect on the model, which may be due to the fact that essays of similar genres have similar essay organization, language expression, etc., but there is no explicit distinction between the different information in the representation of ExtCTS, and the similarity factor will transfer all the information of the essays to the target prompt and thus have a negative effect. Compared to genre information, the prompt unifies the entire essay and the overall representation of the essay is influenced by the prompt information; therefore, using the essay prompt for transfer is effective.

5. Conclusions

In this paper, we propose extra-essay knowledge representation and similarity transfer to enhance shared representation learning in cross-prompt essay trait scoring. Our framework addresses two critical limitations of existing methods: (a) the oversight of external knowledge in measuring inter-prompt commonality, and (b) insufficient alignment between task traits. By explicitly incorporating external knowledge representation to enrich task-specific understanding and quantifying sample similarity through knowledge transfer, we effectively bridge the gap between source and target prompts. Experiments on the ASAP++ dataset show significant improvements over baselines, validating the effectiveness of the proposed method.
In the future, we will further refine the trait descriptions and explore effective combinations of large language models (LLMs), which can give fine-grained evaluations and timely feedback to enhance cross-prompt essay trait scoring. Integrating LLMs into essay scoring systems presents both challenges and opportunities. LLMs might prioritize shallow features (e.g., word choice) over critical thinking, stifling creativity. LLMs also complicate transparent scoring rationales, raising issues of credibility and fairness.
Cross-prompt scoring aims to generalize the evaluation of essays across different prompts or writing tasks, and our model suffers from the following limitations. First, the model is susceptible to prompt bias, and the training data are mostly focused on common prompts, which decreases the accuracy of the evaluation effect for prompts with large domain spans. Second, it is difficult to generalize the scoring criteria, as the core requirements of different writing tasks vary greatly and the scoring scales are different, making it difficult for a single model to accurately adapt to diverse scoring dimensions. Finally, data sparsity limits the robustness of the model, and insufficient samples for some topics may lead to overfitting or underfitting of the scoring rules.

Author Contributions

Conceptualization and methodology, J.S.; writing—original draft preparation, J.S.; writing—review and editing, T.S. and W.P.; supervision, J.S.; funding acquisition, T.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Major Program of National Social Science Foundation of China (No. 22&ZD035), and International Chinese Language Education Research Program (No. 23YH35C).

Data Availability Statement

Publicly available datasets were used in this study. These data can be found here: https://www.kaggle.com/c/asap-aes (accessed on 1 May 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AESAutomated essay scoring
CPETSCross-prompt essay trait scoring
TDETrait Description Encoder
EPEEssay Prompt Encoder

References

  1. Chen, Y.; Li, X. PMAES: Prompt-mapping contrastive learning for cross-prompt automated essay scoring. In Proceedings of the ACL, Toronto, ON, Canada, 9–14 July 2023; pp. 1489–1503. [Google Scholar]
  2. Ke, Z.; Ng, V. Automated essay scoring: A survey of the state of the srt. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; Volume 19, pp. 6300–6308. [Google Scholar]
  3. Uto, M.; Okano, M. Learning automated essay scoring models using item-response-theory-based scores to decrease effects of rater biases. IEEE Trans. Learn. Technol. 2021, 14, 763–776. [Google Scholar] [CrossRef]
  4. Dong, F.; Zhang, Y.; Yang, J. Attention-based recurrent convolutional neural network for automatic essay scoring. In Proceedings of the CoNLL, Vancouver, BC, Canada, 3–4 August 2017; pp. 153–162. [Google Scholar]
  5. Page, E.B. The imminence of grading essays by computer. Phi Delta Kappan 1966, 47, 238–243. [Google Scholar]
  6. Taghipour, K.; Ng, H.T. A neural approach to automated essay scoring. In Proceedings of the EMNLP, Austin, TX, USA, 1–5 November 2016; pp. 1882–1891. [Google Scholar]
  7. Liang, G.; On, B.W.; Jeong, D.; Kim, H.C.; Choi, G.S. Automated essay scoring: A siamese bidirectional LSTM neural network architecture. Symmetry 2018, 10, 682. [Google Scholar] [CrossRef]
  8. Sun, J.; Song, T.; Song, J.; Peng, W. Improving automated essay scoring by prompt prediction and matching. Entropy 2022, 24, 1206. [Google Scholar] [CrossRef] [PubMed]
  9. Higgins, D.; Burstein, J.; Marcu, D.; Gentile, C. Evaluating multiple aspects of coherence in student essays. In Proceedings of the NAACL-HLT, Boston, MA, USA, 2–7 May 2004; pp. 185–192. [Google Scholar]
  10. Mathias, S.; Bhattacharyya, P. Can neural networks automatically score essay traits? In Proceedings of the BEA, Online, 10 July 2020; pp. 85–91. [Google Scholar]
  11. Phandi, P.; Chai, K.M.A.; Ng, H.T. Flexible domain adaptation for automated essay scoring using correlated linear regression. In Proceedings of the EMNLP, Lisbon, Portugal, 17–21 September 2015; pp. 431–439. [Google Scholar]
  12. Ridley, R.; He, L.; Dai, X.y.; Huang, S.; Chen, J. Automated cross-prompt scoring of essay traits. In Proceedings of the AAAI, Palo Alto, CA, USA, 22–24 March 2021; Volume 35, pp. 13745–13753. [Google Scholar]
  13. Xu, J.; Liu, J.; Lin, M.; Lin, J.; Yu, S.; Zhao, L.; Shen, J. EPCTS: Enhanced prompt-aware cross-prompt essay trait scoring. Neurocomputing 2025, 621, 129283. [Google Scholar] [CrossRef]
  14. Wang, J.; Zhang, Q.; Liu, J.; Wang, X.; Xu, M.; Yang, L.; Zhou, J. Making meta-learning solve cross-prompt automatic essay scoring. Expert Syst. Appl. 2025, 272, 126710. [Google Scholar] [CrossRef]
  15. Wang, J.; Liu, J. T-MES: Trait-aware mix-of-experts representation learning for multi-trait essay scoring. In Proceedings of the COLING, Abu Dhabi, United Arab Emirates, 19–24 January 2025; pp. 1224–1236. [Google Scholar]
  16. Zhang, C.; Deng, J.; Dong, X.; Zhao, H.; Liu, K.; Cui, C. Pairwise dual-level alignment for cross-prompt automated essay scoring. Expert Syst. Appl. 2025, 265, 125924. [Google Scholar] [CrossRef]
  17. Ridley, R.; He, L.; Dai, X.; Huang, S.; Chen, J. Prompt agnostic essay scorer: A domain generalization approach to cross-prompt automated essay scoring. arXiv 2020, arXiv:2008.01441. [Google Scholar]
  18. Sun, J.; Peng, W.; Song, T.; Liu, H.; Zhu, S.; Song, J. Enhanced cross-prompt trait scoring via syntactic feature fusion and contrastive learning. J. Supercomput. 2024, 80, 5390–5407. [Google Scholar] [CrossRef]
  19. Do, H.; Kim, Y.; Lee, G.G. Prompt- and Trait Relation-aware cross-prompt essay trait scoring. In Proceedings of the ACL, Toronto, ON, Canada, 9–17 July 2023; pp. 1538–1551. [Google Scholar]
  20. Wu, H.; Huang, D.; Lin, X. Semantic textual similarity with constituent parsing heterogeneous graph attention networks. Symmetry 2025, 17, 486. [Google Scholar] [CrossRef]
  21. Mathias, S.; Bhattacharyya, P. ASAP++: Enriching the ASAP automated essay grading dataset with essay attribute scores. In Proceedings of the LREC, Miyazaki, Japan, 7–12 May 2018; pp. 1169–1173. [Google Scholar]
  22. Dong, F.; Zhang, Y. Automatic features for essay scoring–An empirical study. In Proceedings of the EMNLP, Austin, TX, USA, 1–5 November 2016; pp. 1072–1077. [Google Scholar]
  23. Cozma, M.; Butnaru, A.; Ionescu, R.T. Automated essay scoring with string kernels and word embeddings. In Proceedings of the ACL, Melbourne, Australia, 15–20 July 2018; pp. 503–509. [Google Scholar]
  24. Jin, C.; He, B.; Hui, K.; Sun, L. TDNN: A two-stage deep neural network for prompt-independent automated essay scoring. In Proceedings of the ACL, Melbourne, Australia, 15–20 July 2018; pp. 1088–1097. [Google Scholar]
  25. Song, W.; Zhang, K.; Fu, R.; Liu, L.; Liu, T.; Cheng, M. Multi-stage pre-training for automated Chinese essay scoring. In Proceedings of the EMNLP, Online, 16–20 November 2020; pp. 6723–6733. [Google Scholar]
  26. Cao, Y.; Jin, H.; Wan, X.; Yu, Z. Domain-adaptive neural automated essay scoring. In Proceedings of the SIGIR, Xi’an, China, 25–30 July 2020; pp. 1011–1020. [Google Scholar]
  27. Jiang, Z.; Gao, T.; Yin, Y.; Liu, M.; Yu, H.; Cheng, Z.; Gu, Q. Improving domain generalization for prompt-aware essay scoring via disentangled representation learning. In Proceedings of the ACL, Toronto, ON, Canada, 9–14 July 2023; pp. 12456–12470. [Google Scholar]
  28. Li, X.; Chen, M.; Nie, J.Y. SEDNN: Shared and enhanced deep neural network model for cross-prompt automated essay scoring. Knowl.-Based Syst. 2020, 210, 106491. [Google Scholar] [CrossRef]
  29. Li, J.; Fukumoto, F. Multi-task neural shared structure search: A study based on text mining. In Proceedings of the DASFAA, Taipei, Taiwan, 11–14 April 2021; pp. 202–218. [Google Scholar]
  30. Banerjee, S.; Akkaya, C.; Perez-Sorrosal, F.; Tsioutsiouliklis, K. Hierarchical transfer learning for multi-label text classification. In Proceedings of the ACL, Florence, Italy, 28 July–2 August 2019; pp. 6295–6300. [Google Scholar]
  31. Chen, C.; Huang, X.; Ji, F.; Wang, C.; Qiu, M.; Huang, J.; Zhang, Y. CAT-BERT: A context-aware transferable BERT model for multi-turn machine reading comprehension. In Proceedings of the DASFAA, Taipei, Taiwan, 11–14 April 2021; pp. 152–167. [Google Scholar]
  32. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the EMNLP, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  33. Hussein, M.A.; Hassan, H.A.; Nassef, M. A trait-based deep learning automated essay scoring system with adaptive feedback. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 287–293. [Google Scholar] [CrossRef]
  34. Li, S.; Ng, V. Conundrums in Ccross-prompt automated essay scoring: Making sense of the state of the srt. In Proceedings of the ACL, Bangkok, Thailand, 11–16 August 2024; pp. 7661–7681. [Google Scholar]
  35. Uto, M.; Xie, Y.; Ueno, M. Neural automated essay scoring incorporating handcrafted features. In Proceedings of the COLING, Barcelona, Spain, 8–13 December 2020; pp. 6077–6088. [Google Scholar]
  36. Chen, Y.; Li, X. PLAES: Prompt-generalized and level-aware learning framework for cross-prompt automated essay scoring. In Proceedings of the LREC-COLING, Torino, Italy, 20–24 May 2024; pp. 12775–12786. [Google Scholar]
Figure 1. Settings of holistic scoring and trait scoring across cross-prompt scenarios.
Figure 1. Settings of holistic scoring and trait scoring across cross-prompt scenarios.
Symmetry 17 00739 g001
Figure 2. The ExtCTS model.
Figure 2. The ExtCTS model.
Symmetry 17 00739 g002
Figure 3. The encoder modules. Subfigures (a,b) are the submodules of the Trait Description Encoder and the Essay Prompt Encoder, respectively, and subfigures (c,d) are the submodules of the Trait-Specific Encoder.
Figure 3. The encoder modules. Subfigures (a,b) are the submodules of the Trait Description Encoder and the Essay Prompt Encoder, respectively, and subfigures (c,d) are the submodules of the Trait-Specific Encoder.
Symmetry 17 00739 g003
Figure 4. The effect of different trait similarity loss functions for different prompts.
Figure 4. The effect of different trait similarity loss functions for different prompts.
Symmetry 17 00739 g004
Figure 5. The effect of different trait similarity loss functions for different traits.
Figure 5. The effect of different trait similarity loss functions for different traits.
Symmetry 17 00739 g005
Table 1. Examples of prompts and partial essay texts.
Table 1. Examples of prompts and partial essay texts.
Prompt ID (Genre): Prompt Text
Prompt 1 (argumentative): Write a letter to your local newspaper in which you state your opinion on the effects computers have on people. Persuade the readers to agree with you.
Prompt 2 (argumentative): Write a persuasive essay to a newspaper reflecting your vies on censorship in libraries.
Prompt 3 (narrative): Tell a true story in which laughter was one element or part.
Prompt 1 example: Censorship is found on television, in libraries, in schools, and even in homes. Do you think that censorship should be used? Censorship can be used to block one’s personal thoughts and opinions to promote conformism. I disagree with the idea of Censorship. Not everyone thinks people should be able to have their own images and visions; however, that is what makes our world so individualized.
Prompt 2 example: Final exams, omnipresent in schools, are divisive. Do they truly benefit students? I argue against their overemphasis. They can overshadow holistic learning, promoting stress over understanding. Yet, it’s this very pressure that ignites the drive for mastery, shaping individual academic journeys.
Prompt 3 example: Laughter is a really good thing to have in any relationship you are in whether its a friend ship, a love relation ship, or work relation ship. Either way you need laughter so you don’t get so up tight that you are no fun to be around. Every now and then some body around you in your life needs some laughter to cheer them up.
Table 2. ASAP++ dataset statistics for cross-prompt scoring.
Table 2. ASAP++ dataset statistics for cross-prompt scoring.
PromptNo. EssayMax LenAvg LenTraitsGenreRange
11783950350OA, Cont, Org, WC, SF, ConvARG[2, 12]
218001173350OA, Cont, Org, WC, SF, ConvARG[1, 6]
31726415150OA, Cont, PA, Lang, NarRES[0, 3]
41772395150OA, Cont, PA, Lang, NarRES[0, 3]
51805474150OA, Cont, PA, Lang, NarRES[0, 4]
61800502150OA, Cont, PA, Lang, NarRES[0, 4]
71569732250OA, Cont, Org, ConvNAR[0, 30]
87231146650OA, Cont, Org, WC, SF, ConvNAR[0, 60]
ARG: Argumentative; RES: Source-Dependent Response; NAR: Narrative; OA: Overall; Cont: Content; Org: Organization; WC: Word Choice; SF: Sentence Fluency; Conv: Conventions; PA: Prompt Adherence; Lang: Language; Nar: Narrativity.
Table 3. Description of the trait of the ASAP++ dataset [21].
Table 3. Description of the trait of the ASAP++ dataset [21].
Trait: description
Genre: Argumentative, Narrative
Content: The amount of relevant text present in the essay.
Organization: The way the essay is structured.
Word Choice: The choice and aptness of the vocabulary used in the essay.
Sentence Fluency: The quality of the sentences in the essay.
Conventions: Overall writing conventions to be followed, like spelling, punctuations, etc.
Style: How well the essay is written consists of two major parts, namely sentence fluency and word choice.
Voice: The quality and accuracy of voice.
Genre: Source-Dependent Response
Content: The amount of relevant text present in the essay.
Prompt Adherence: A measure of how the writer sticks to the question asked in the prompt.
Language: The quality of the grammar and spelling in the response.
Narrativity: A measure of the coherence and cohesion of the response to the prompt.
Table 4. Description of the genre of the ASAP++ dataset [21].
Table 4. Description of the genre of the ASAP++ dataset [21].
Genre: description
Argumentative: These are essays where the prompt is one in which the writer has to convince the reader about their stance for or against a topic (for example, free speech in public colleges).
Source-Dependent Response: These essays are responses to a source text, where the writer responds to a question about the text (for instance, describing the writer’s opinion about an incident that happened to him in the text).
Narrative: These are essays where the prompt requires us to describe or narrate a story.
Table 5. Average results for different traits.
Table 5. Average results for different traits.
ModelPromptsAVG QWK ↑
12345678
Hi att 0.3150.4780.3170.4780.3750.3570.2050.2650.349
AES aug 0.3300.5180.2990.4770.3410.3990.1620.2000.341
PAES 0.6050.5220.5750.6060.6340.5450.3560.4470.536
CTS w/o att0.6190.5390.5850.6160.6160.5440.3630.4610.543
CTS0.6230.5400.5920.6230.6130.5480.3840.5040.553
PMAES0.6560.5530.5980.6060.6260.5720.3860.5300.566
RDCTS0.6510.5530.6080.6230.6510.5800.3750.5290.571
PLAES0.6480.5630.6040.6230.6340.5930.4030.5330.575
ProTACT0.6470.5870.6230.6320.6740.5840.4460.5410.592
ExtCTS ( BERT b a s e )0.6650.5830.6460.6560.6750.6110.4210.5620.602
ExtCTS ( BERT l a r g e )0.6700.5960.6480.6540.6740.6030.4220.5430.601
The best measures are in bold; : results from [12]. The upperarrow indicates that a higher value is better.
Table 6. Average results for different traits.
Table 6. Average results for different traits.
ModelTraitsAVG QWK ↑
OAContOrgWCSFConvPALangNar
Hi att 0.4530.3480.2430.4160.4280.2440.3090.2930.3790.346
AES aug 0.4020.3420.2560.4020.4320.2390.3310.3130.3770.344
PAES 0.6570.5390.4140.5310.5360.3570.5700.5310.6050.527
CTS w/o att0.6590.5410.4240.5580.5440.3870.5610.5390.6050.535
CTS0.6700.5550.4580.5570.5450.4120.5650.5360.6080.545
FEATS0.6820.5920.4780.4590.4520.4390.6170.5560.6370.546
PMAES0.6710.5670.4810.5840.5820.4210.5840.5450.6140.561
RDCTS0.6730.5610.4800.5910.5760.4260.6090.5600.6340.568
PLAES0.6730.5740.4910.5790.5800.4470.6010.5540.6310.570
ProTACT0.6740.5960.5180.5990.5850.4500.6190.5960.6390.586
ExtCTS ( BERT b a s e )0.6810.6120.5210.6170.6040.4510.6350.6070.6540.598
ExtCTS ( BERT l a r g e )0.6830.6080.5110.6190.6100.4580.6330.6060.6510.598
OA: Overall; Cont: Content; Org: Organization; WC: Word Choice; SF: Sentence Fluency; Conv: Conventions; PA: Prompt Adherence; Lang: Language; Nar: Narrativity. The best measures are in bold; : results from [12]. The upperarrow indicates that a higher value is better.
Table 7. Ablation results.
Table 7. Ablation results.
NO.ModelAVG QWK ↑
1ExtCTS0.602
2   w/o prompt representation h 0.598 (−0.004)
3   w/o trait description representation d 0.600 (−0.002)
4   w/o target prompt aware representation s 0.595 (−0.007)
5   w/o all extra-essay knowledge representation0.593 (−0.009)
6   w/o shallow level feature transfer0.604 (+0.002)
7   w/o middle level feature transfer0.597 (−0.005)
8   w/o deep level feature transfer0.598 (−0.004)
9   w/o shallow and middle level feature transfer0.598 (−0.003)
10   w/o shallow and deep level feature transfer0.596 (−0.006)
11   w/o middle and deep level feature transfer0.590 (−0.012)
The best measures are in bold. The upperarrow indicates that a higher value is better.
Table 8. Effect of similarity loss function.
Table 8. Effect of similarity loss function.
η PromptAvg QWK ↑
12345678
10.6350.5800.6430.6420.6770.6070.4070.5220.589
0.70.6650.5830.6460.6560.6750.6110.4210.5620.602
The best measures are in bold. The upperarrow indicates that a higher value is better.
Table 9. The effect of different transfer factors.
Table 9. The effect of different transfer factors.
f sim PromptsAVG QWK ↑
12345678
f p 0.6650.5830.6460.6560.6750.6110.4210.5620.602
f g 0.6560.5880.6300.6410.6520.6000.4050.4050.572
The best measures are in bold. The upperarrow indicates that a higher value is better.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, T.; Sun, J.; Peng, W. Enhancing Cross-Prompt Essay Trait Scoring via External Knowledge Similarity Transfer. Symmetry 2025, 17, 739. https://doi.org/10.3390/sym17050739

AMA Style

Song T, Sun J, Peng W. Enhancing Cross-Prompt Essay Trait Scoring via External Knowledge Similarity Transfer. Symmetry. 2025; 17(5):739. https://doi.org/10.3390/sym17050739

Chicago/Turabian Style

Song, Tianbao, Jingbo Sun, and Weiming Peng. 2025. "Enhancing Cross-Prompt Essay Trait Scoring via External Knowledge Similarity Transfer" Symmetry 17, no. 5: 739. https://doi.org/10.3390/sym17050739

APA Style

Song, T., Sun, J., & Peng, W. (2025). Enhancing Cross-Prompt Essay Trait Scoring via External Knowledge Similarity Transfer. Symmetry, 17(5), 739. https://doi.org/10.3390/sym17050739

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop