Next Article in Journal
Addressing Challenges in Rds,on Measurement for Cloud-Connected Condition Monitoring in WBG Power Converter Applications
Previous Article in Journal
A Novel Harmonic Clocking Scheme for Concurrent N-Path Reception in Wireless and GNSS Applications
Previous Article in Special Issue
Analysis of Student Progression Through Curricular Networks: A Case Study in an Illinois Public Institution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quality over Quantity: An Effective Large-Scale Data Reduction Strategy Based on Pointwise V-Information

Information Science and Technology College, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(15), 3092; https://doi.org/10.3390/electronics14153092
Submission received: 15 June 2025 / Revised: 29 July 2025 / Accepted: 31 July 2025 / Published: 1 August 2025
(This article belongs to the Special Issue Data Retrieval and Data Mining)

Abstract

In order to increase the effectiveness of model training, data reduction is essential to data-centric Artificial Intelligence (AI). It achieves this by locating the most instructive examples in massive datasets. To increase data quality and training efficiency, the main difficulty is choosing the best examples rather than the complete datasets. In this paper, we propose an effective data reduction strategy based on Pointwise V -Information (PVI). To enable a static method, we first use PVI to quantify instance difficulty and remove instances with low difficulty. Experiments show that classifier performance is maintained with only a 0.0001% to 0.76% decline in accuracy when 10–30% of the data is removed. Second, we train the classifiers using a progressive learning strategy on examples sorted by increasing PVI, accelerating convergence and achieving a 0.8% accuracy gain over conventional training. Our findings imply that training a classifier on the chosen optimal subset may improve model performance and increase training efficiency when combined with an efficient data reduction strategy. Furthermore, we have adapted the PVI framework, which was previously limited to English datasets, to a variety of Chinese Natural Language Processing (NLP) tasks and base models, yielding insightful results for faster training and cross-lingual data reduction.

1. Introduction

Driven by large-scale datasets, large language models, pre-training, and finetuning training procedures, Artificial Intelligence (AI) technology has made remarkable progress in the field of Natural Language Processing (NLP). With the widespread application of AI systems, it has become increasingly clear that model performance is greatly influenced by the quality of data [1]. Traditionally, the Model-Centric paradigm has been used to conduct AI research, with a primary focus on developing new model architectures and suggesting optimal algorithms to boost performance [2]. However, this paradigm often overlooks the intrinsic quality of data. Issues including imbalance, labeling errors, and redundant data may result in deteriorated model performance, skewed outcomes, and poor decision-making [3,4,5]. As they say, “garbage in, garbage out.” It is more beneficial to improve the quality of data than to merely increase its quantity [6,7]. As a result, a new paradigm known as data-centric AI has surfaced, emphasizing systematic improvement in data quality to enhance model performance using fewer but high-quality data. The importance of the data-centric paradigm is being recognized, advocating a shift in focus from continuously improving model architectures and algorithm optimization to prioritizing the enhancement of high-quality data [8].
Data reduction is a fundamental strategy in data-centric AI [9], which assesses data quality, eliminates low-quality data, and retains predominantly high-quality data in a dataset to optimize model training efficacy and performance. One of the representative methods for dataset reduction is dataset distillation, which aims to improve data processing efficiency by synthesizing a small typical dataset from substantial data [10]. The training of large language models often relies on large-scale datasets [11], which may diminish training efficiency and model generalization due to the presence of redundant low-quality data and incur significant computational and storage costs. To reduce the training dataset to a manageable size while maintaining model performance, the primary challenge in data reduction is selecting the optimal subset from large-scale datasets utilizing data quality measures as a reference.
The term dataset difficulty pertains to the quality of the data. It is fundamentally rooted in information theory and serves in the generalization of mutual information. It assesses the informational richness or learning complexity of data. The assumption of dataset difficulty is that complex knowledge and challenging data contribute to the development of powerful models whereas an abundance of low-quality data impedes the model learning efficiency.
Several measures have been proposed as potential indicators of a dataset’s difficulty. Devin Kwok [12] focused on example difficulty scores, such as Prediction Depth [13] and Variance of Gradients (VoG) [14]. Peng Cui et al. [15] evaluated sample difficulty by employing feature-space Gaussian modeling and relative Martens distance calculation. David Mayo et al. [16] introduced Minimum Viewing Time as a dataset difficulty measure. Chengwen Wang et al. [17] proposed four difficulty measures to be applied to named entity recognition datasets, including three internal measures (invisible entity ratio, entity ambiguity, and text complexity) and one external measure (model variance).
Pointwise V -Information (PVI) [18] is a promising metric for quantifying dataset difficulty which defines dataset difficulty as the lack of model-usable information. After processing the original data X , the increase in the ability of humans or machine learning algorithms to predict the label Y is the usable information. More formally, let t be the decryption algorithm and V be a class of processing functions. It follows that I V t X Y > I V ( X Y ) 0 . The less usable information there is, the more difficult the dataset is for a model. Furthermore, PVI could measure the difficulty of each instance in a given dataset. Within this framework, instance difficulty is defined as the inherent challenge that an individual data point presents to a specific model in learning or prediction. The instance difficulty is inversely proportional to the amount of V -usable Information the instance provides. There is a difference between instance difficulty and instance usefulness; the latter should be the contribution an individual data instance makes to a model’s generalization capability and overall performance. An instance’s usefulness is a multifaceted concept that depends on both its inherent difficulty and its relevance to the specific learning task. For data instances, high PVI indicates that they are easy for the model to learn. During training, a small number of easy instances can elevate model performance to a certain level, but continuously feeding easy instances yields minimal performance gains. Low PVI indicates that the instances are hard to learn but could gain a larger margin of performance than the easy instances. This suggests that excessive easy instances, marked by high PVI, are redundant, leading to a significant waste of computational resources and disproportional performance gains.
The PVI framework provides new perspectives on dataset evaluation, selection, and reduction. However, since most research has been conducted on English datasets, there were concerns about its applications in a cross-lingual context. Is it possible to generalize the data reduction approach and dataset difficulty measures to other languages? How can we use PVI to boost model performance in a cross-lingual context and increase training efficiency?
To address these issues, we present a PVI-based large-scale data reduction strategy. Our research focuses on obtaining an optimal subset for training while improving training efficiencies, saving computational resources, and preserving model performance. The contributions of our paper are as follows:
  • We eliminated low-quality instances by employing PVI to assess instance difficulty. Experiments demonstrate that the removal of 10–30% of the data results in a minimal decline in classifier performance, ranging from 0.0001% to 0.76% in accuracy. This indicates that we might effectively sustain model performance and accelerate training by removing a certain quantity of low-quality instances.
  • We proposed a progressive learning technique that trained a classifier by initially organizing examples in descending order of PVI. This challenging training method not only accelerated model convergence but also resulted in a 0.8% enhancement in accuracy. The results indicate that strategically employing PVI to guide the training process could significantly enhance model performance and training efficiency. In contrast to current curriculum learning methods that depend on heuristic or less universal difficulty metrics, our approach utilizes the robust and adaptable characteristics of PVI to effectively direct the learning process, resulting in enhanced training efficiency and superior generalization for large language models.
  • The PVI framework, hitherto restricted to English datasets, has been adapted for various Chinese NLP tasks and fundamental models. The cross-lingual extension provides novel perspectives for data-centric AI in broader application contexts, confirms the universality of the PVI framework, and supplies valuable insights for cross-lingual data reduction.

2. Materials and Methods

Alex Havrilla [19] notes, in his definition of dataset complexity, that complexity is a data characteristic that intuitively reflects the difficulty of a sample. He defines the complexity C ( ω ) of an instance ω Ω as its size under a fixed representation scheme. An n-sample complexity measure is represented as a function C   :     Ω n R , which intuitively measures the difficulty of the data, defining complexity C Ω R at the level of a single sample, with C being recovered as the average over samples.
Several fixed representation schemes mentioned above exhibit certain limitations. For instance, the example difficulty scores used by Devin Kwok include various scores for quantifying the difficulty of individual instances in the training dataset, which typically depend on the model. The relative Martens distance calculation used by Peng Cui et al. is primarily applied in computer vision tasks such as image classification. The Minimum Viewing Time introduced by David Mayo et al. is also limited to quantifying the difficulty of computer vision datasets [20,21]. In NLP, the entity ratio, entity ambiguity, text complexity, and model variance used by Chengwen Wang et al. are only targeted at named entity recognition datasets [22,23].
In contrast, PVI offers a more universal and flexible approach to measuring the instance difficulty. PVI quantifies the difficulty of the individual instance within a given distribution, framing dataset difficulty with respect to a model V . Dataset difficulty is conceptualized as the lack of information readily usable by model V . A significant advantage of PVI is its ability to facilitate cross-dataset difficulty comparisons, even across diverse label spaces. This inherent flexibility provides PVI with a much broader application scope compared to the traditional performance metrics.
The escalating scale of modern datasets poses substantial challenges for model training, which necessitates immense computational resources and prolonged training times. While existing methods offer valuable insights into dataset difficulty, their inherent limitations often restrict their applicability to diverse and large-scale scenarios. PVI addresses these challenges by providing a model-aware quantification of individual instance difficulty. By identifying and prioritizing data instances based on their PVI, we can significantly streamline the training process, and enable data reduction while maintaining model performance.

2.1. Model Architecture

The purpose of this paper is to construct an efficient data reduction strategy to optimize the efficiency of data usage in natural language inference (NLI) [24] tasks by quantifying and using the difficulty of data instances. To this end, we designed a comprehensive framework that includes three modules: data transformation, PVI calculator, and reduction approach. The overall architecture is shown in Figure 1.
Data Transformation: This module is the preprocessing stage of the entire process and is responsible for converting the original dataset into a variety of input formats required by subsequent modules. It is an NLI Transformation base class, which defines standard processes for data loading, filtering, and preservation. We have obtained various data transformation results, the two most important of which are as follows: standard input, a standard NLI input containing prerequisites and hypotheses as input features ( X ); and null input, an empty string ( ) that does not provide any information and is essential for calculating the model prior predictive ability in the absence of explicit evidence.
PVI Calculator: This module is responsible for calculating the V -entropy and PVI of the dataset to quantify the amount of information in each data instance. The computational process for PVI and V -information is shown in Algorithm 1. PVI measures the gain in the model predictive confidence in the correct label y after receiving the standard input x compared to receiving null input. According to the definition of Xu et al. [25], let X and Y represent random variables with instance space X and Y , respectively. Let represent an empty input that does not provide information about Y .
Let Ω = { f : X { } P ( Y ) } . A predictive family is a set of predictive models that the agent is allowed to use; we say that V Ω is a predictive family if it satisfies
f V ,   P r a n g e f ,   f V ,   s . t .   x X ,   f x = P ,   f = P ,
Given the predictive family, the predicted V -entropy is
H V Y = E f V i n f [ l o g 2 f [ ] ( Y ) ] ,
and the conditional V -entropy is
H V Y | X = E f V i n f [ l o g 2 f [ X ] ( Y ) ] ,
where l o g 2 is used to measure the entropy of information bits.
I V X Y = H V Y H V ( Y | X ) ,
Shannon’s mutual information I ( X ; Y ) serves as a direct analog to accessible information [26]. It quantifies the amount of information obtained about one random variable by observing another. I ( X ; Y ) measures the reduction in uncertainty about Y given X , assuming an observer with unbounded computational power. Information can be “accessible” in a statistical sense but not practically extractable or actionable by a computationally constrained agent. Useful information is the subset of accessible information that a specific, computationally constrained observer or model V can actually extract, process, and leverage to perform a given task. It is the information that is actionable and predictive for a defined computational agent. PVI is built on the theory of V -information [25], with the V -information I V ( X Y ) in Formula (4) being the difference between the V -entropy H V ( Y ) and the conditional V -entropy H V ( Y | X ) . I V ( X Y ) reflects how much the model V can reduce its uncertainty about Y given X . The V -entropy measures the uncertainty of the model in predicting labels without input, while the conditional V -entropy measures the uncertainty with input X . High PVI indicates that the instance is “simpler” for the model, as input x provides more effective information for the correct prediction of y . Therefore, an instance ( x , y ) is defined as a simple instance if P V I ( x y ) > τ , where τ is a threshold determined based on the specific task and model performance. Let D be a dataset; a subset D e a s y D containing many simple instances is considered redundant if it satisfies the following condition:
lim D ea sy D easy P e r f o r m a n c e ( V , D ) P e r f o r m a n c e ( V , D D e a s y ) ϵ p e r f ,
After the model V is trained on D and D D e a s y separately, the performance gap P e r f o r m a n c e (e.g., accuracy, F1 score on the validation or test set) resulting from removing more simple instances D e a s y D e a s y from the training data tends to an acceptable minimum value ϵ p e r f . The acceptable removal rate for D e a s y adapts to different tasks and datasets, a principle incorporated into the reduction ratios r of Algorithms 2 and 3. Consequently, changes to D e a s y directly affect the model’s performance(see Section 3.2 for details).
According to the definition of Kawin Ethayarajh [18], the calculation formula for the PVI of an instance ( x ,   y ) is as follows:
P V I x y = l o g 2 g y + l o g 2 g x y ,
where g and g are the models selected from the predictive family V ; for example, they can be BERT-family models finetuned under both standard input ( x ) and null input ( ) . Actually, according to Xu et al.’s definition [25], the V is a mapping set from the input space to the output probability distribution; i.e., V { f : X P ( Y ) } , where P   ( Y ) is the probability distribution set on the output space Y . g usually corresponds to the best predictive function that the model can achieve in the V family given the input X . g represents the best possible predictor for output Y without a specific input X . g [ x ] ( y ) is the logarithm of the probability that the model predicts y as the correct label after seeing standard input x . g [ ] ( y ) is the logarithm of the probability that the model predicts y as the correct label seeing the null input.
The PVI calculator module receives standard input and null input datasets generated by the data transformation module, along with a pre-trained text classification model (e.g., Chinese-BERT-wwm [27], BERT-base-Chinese [28], and Chinese-MacBERT [29]) and the tokenizer. For each instance, the module calculates its log-likelihood H y b corresponding to l o g 2 g [ x ] ( y ) and H y x corresponding to l o g 2 g [ ] ( y ) in Formula (6) for standard and null inputs. Finally, the module outputs a series of quantified metrics for each instance, including the PVI, and sorts the instances in the dataset by PVI.
Reduction Approach: After obtaining the PVI for all training instances, this module is responsible for conducting the data reduction strategies and evaluating their effectiveness. We have designed two reduction methods, implemented, respectively, by Algorithm 2 and Algorithm 3 (see Section 2.2 for details). Static reduction (Algorithm 2) is a method that aims to evaluate the value of difficult instances. It filters out the subset of instances with low PVI based on a reduction ratio r , trains the Chinese-BERT-wwm model from scratch using the subset, and finally evaluates its accuracy on the test set. Progressive learning (Algorithm 3) is a method that adopts a strategy similar to curriculum learning [30], designed to improve training efficiency. It first allows the model to learn from simple instances with high PVI, and then gradually introduces more difficult instances. Finally, it evaluates the accuracy, precision, recall, and F1 score on the test set.
We utilize cross-entropy loss [31] as the optimization objective for model training. Specifically, for a training batch containing N instances, the loss function   J ( θ ) is defined as follows:
J θ = L b a t c h = 1 N i = 1 N c = 1 C y i c l o g ( y ^ i c ) ,
where θ represents the trainable parameters of the model. N is the number of instances in the current training batch. C is the total number of categories, and in an NLI task, C equals 3. y ^ i c is the probability that the model predicts that the i th instance belongs to category C .
The training objective of the model is to find a set of parameters θ that minimizes the value of the loss function J ( θ ) :
arg  m in θ J ( θ )

2.2. Algorithm

Algorithm 1 is the computational process for PVI and V -information. The PVI measures the instance difficulty by comparing the change in confidence of the model predictions for an instance. An instance with high PVI is typically easy to predict, whereas low PVI indicates that the model finds inference for the instance more challenging.
Algorithm 1 calculates the total amount of information provided by input features to the prediction of the target label from the model’s perspective. By comparing the predictive capabilities of g and g , the algorithm can analyze the gain of input feature x i in correctly predicting the model’s label. g represents the baseline predictive capability of the model without input features, while g represents the model’s predictive capability conditioned on input features.
Algorithm 1: PVI Calculator After finetuning on a dataset of size n, the V -information and PVI can be calculated in O ( n ) time
Input: training data D t r a i n = i n p u t   x i , g o l d   l a b e l   y i i = 1 m , held-out data D t e s t = i n p u t   x i , g o l d   l a b e l   y i i = 1 n , model V
do
   g  ← Finetune V on D t r a i n
   ← empty string (null input)
   g ← Finetune V on { ( ,   y i ) | ( x i , y i ) D t r a i n }
   H V Y , H V Y | X 0 , 0
  for  ( x i , y i ) D t e s t  do
       H V Y H V Y 1 n l o g 2 g [ ] ( y i )
       H V Y | X H V Y | X 1 n l o g 2 g [ x i ] ( y i )
       P V I x i y i l o g 2 g y i + l o g 2 g [ x i ] ( y i )
   end for
   I ^ V X Y = 1 n i P V I x i y i = H V Y H V Y | X
end do
We quantify the amount of information that different datasets provide to the model. Figure 2 illustrates the results of different Chinese datasets providing varying amounts of information to the same model, Chinese-BERT-wwm. According to the distribution of dataset difficulty, the OCNLI dataset contains more information usable by Chinese-BERT-wwm compared to the CMNLI and CINLI datasets, making the computation based on Chinese-BERT-wwm easier.
Algorithm 2 aims to investigate the relationship between the difficulty of training instances and the model performance through a static data reduction method. Its objective is to evaluate the necessity or redundancy of the simple instances during the model training process and to validate a hypothesis: training the model exclusively with the instances deemed difficult by the model can effectively enhance its generalization ability. The algorithm employs a static strategy, meaning that each experiment uses a fixed, preselected data subset based on a specific difficulty threshold to train a completely new model from scratch. Algorithm 2 first performs PVI computation and difficulty sorting on the entire dataset, using a model finetuned on the full training set. With this model, it calculates the corresponding PVI for each instance ( x i ,   y i )   in D t r a i n . After computation, the entire training set is sorted in descending order based on PVI, so that the simple instances with high PVI are at the head of the list, while the difficult instances with low PVI are at the tail. m represents the total number of instances in the training dataset D t r a i n . Algorithm 2 is a cyclic process that iterates through a series of reduction ratios r (from 0.1 to 0.9). In each iteration, the subset size to be retained is calculated based on the reduction ratio r . As r increases, s u b s e t s i z e decreases accordingly, meaning that the selected subset contains fewer instances but high average difficulty. To maintain the original batch processing order during training, the selected subset is reordered based on its original indices to obtain the training subset. For each difficult data subset D s u b s e t generated through different reduction ratios r , the algorithm initializes a completely new, untrained model m o d e l r , which is finetuned exclusively using the corresponding D s u b s e t . After training, the accuracy D s u b s e t of m o d e l r is evaluated on the held-out test set D t e s t , and the performance under this reduction ratio is recorded (see Section 3.2.1 for details).
Algorithm 2: Static reduction PVI-based static data reduction for accuracy analysis
Input: original training data D t r a i n o r i g i n a l , held-out data D t e s t , model V , pre-prepared imbalanced training data D t r a i n i m b a l a n c e d , pre-prepared noisy training data (noisy level = 0.1) D t r a i n n o i s y
do
    for each  D t r a i n in [original, imbalanced, noisy] do
         g ← Finetune V on D t r a i n
        Calculate P V I ( x i y i )  for all ( x i , y i ) D t r a i n
         D t r a i n   sorted ← Sort D t r a i n instances by PVI in descending order
        for  r  in [0.1, 0.2, …, 0.9] do
                s u b s e t s i z e   m     ( 1     r )
                D s u b s e t   sorted ← Select the last s u b s e t s i z e instances from D t r a i n   sorted
      D s u b s e t ← reorder D s u b s e t   sorted by original_idx_i
                m o d e l r ← Initialize a new model
               Finetune m o d e l r on D s u b s e t
                  Evaluate m o d e l r on D t e s t and record A C C for reduction ratio r
        end for
     end for
end do
Algorithm 3 aims to explore a progressive learning approach, which is inspired by the concept of curriculum learning. The objective of this algorithm is to validate the hypothesis that by carefully arranging the order of the training instances, easy first, then hard, it can optimize the finetuning process of large language models [32] (https://huggingface.co/Qwen/Qwen3-0.6B (accessed on 15 June 2025), Qwen3-0.6B [33]), thereby achieving faster convergence and better generalization.
Unlike Algorithm 2, which analyzes model performance by statically removing data, Algorithm 3 focuses on dynamically and incrementally feeding data to the model. It first utilizes PVI to rank the entire training set in terms of difficulty, then starts training from the simple instances, and gradually expands the training set to include more difficult instances. Algorithm 3 organizes the instances in an ordered manner from simplest (highest PVI) to most difficult (lowest PVI) from the model’s perspective. After each progressive training stage is completed, the model ( m o d e l r ) trained on the data subset of that stage is evaluated on the held-out test set D t e s t . For detailed performance analysis, the evaluation metrics include accuracy, precision, recall, and F1 score. By recording and comparing these metrics at different stages, it becomes clear how model performance evolves as the difficulty and quantity of training data increase (see Section 3.2.2 for details).
Algorithm 3: Progressive learning PVI-based data reduction and progressive learning for detailed performance evaluation
Input: original training data D t r a i n o r i g i n a l , held-out data D t e s t , model V , pre-prepared imbalanced training data D t r a i n i m b a l a n c e d , pre-prepared noisy training data (noisy level = 0.1) D t r a i n n o i s y
do
    for each  D t r a i n in [original, imbalanced, noisy] do
         g ← Finetune V on D t r a i n
        Calculate P V I ( x i y i )  for all ( x i , y i ) D t r a i n
         D t r a i n   sorted ← Sort D t r a i n instances by PVI in descending order
        for  r  in [0,0.1,0.2,0.3] do
             s u b s e t s i z e   m     ( 1     r )
             D s u b s e t ← Select the last s u b s e t s i z e instances from D t r a i n sorted
             m o d e l r ← Initialize a new model
            Finetune m o d e l r on D s u b s e t
    Evaluate m o d e l r on D t e s t and record Accuracy, Precision, Recall, F1 for reduction ratio r
     end for
    end for
end do

3. Experiments and Results

3.1. Experimental Setup

Dataset: We utilized three Chinese natural language inference datasets: OCNLI, CMNLI, and CINLI. All datasets contain premise–hypothesis pairs as input features and are annotated with entailment, contradiction, or neutral labels. OCNLI [34] (Original Chinese Natural Language Inference dataset) contains approximately 56,000 premise–hypothesis pairs, entirely based on original Chinese materials. CMNLI [35] (Chinese Multi-Genre Natural Language Inference dataset) integrates Chinese data from XNLI [36] and MultiNLI [37], covering various genres such as news and fiction, used to evaluate cross-domain NLI capabilities. CINLI (Chinese Idioms Natural Language Inference Dataset) focuses on NLI tasks involving Chinese idioms and colloquialisms, containing 91,247 manually annotated idiom pairs, designed to assess models’ understanding of subtle semantic differences in Chinese. Before the experiments, we preprocessed the datasets, removing corrupted or incorrectly formatted pairs. The statistical information of the datasets used in the experiments is shown in Table 1, which summarizes the scale and label category statistics for each dataset.
Hyperparameter Setting: For the Chinese-BERT-wwm model, the maximum sequence length is set to 128 tokens, ensuring both the integrity of model input and the optimization of computational resource utilization. The batch size is set to 32, enabling good parallel processing capabilities on most common hardware configurations. The learning rate is set to 5 × 10−5, which is a common starting value for finetuning BERT-series models, balancing the model’s convergence speed with final performance. The training period is set to 2 epochs, and a linear learning rate scheduler is selected to effectively manage the dynamic changes in the learning rate. Additionally, the gradient accumulation step is set to 1, with gradient updates performed independently for each batch. To ensure the reproducibility of experimental results, a fixed random seed of 1 is set.
The hyperparameter settings for the Qwen3-0.6B model differ to accommodate its model architecture characteristics. The maximum sequence length is extended to 512 tokens to handle longer context information. The batch size is uniformly set to 8, balancing training efficiency and resource consumption under memory-limited conditions. The learning rate is set to 2 × 10−5, accompanied by a weight decay of 0.01, to achieve more stable training convergence and prevent overfitting. The model is also trained for 2 epochs, with evaluation and saving strategies set to execute after each epoch ends, facilitating periodic monitoring of model performance and saving the best checkpoints. The logging step is set to 50, enabling fine-grained tracking of the training process. To improve training efficiency and reduce GPU memory usage, mixed-precision training (fp16 = True) is enabled.

3.2. Result Analysis

3.2.1. Static Reduction

According to the PVI theory [18], we conducted difficulty analysis and static reduction experiments on the Chinese NLI datasets OCNLI, CMNLI, and CINLI. The theory indicates that high-PVI instances suggest that the model can easily extract information strongly associated with the label y from the input x . These instances may contain annotation artifacts (such as high-frequency words, fixed patterns) or shallow patterns, leading the model to achieve high accuracy through “shortcut learning” rather than deep semantic inference. Therefore, removing such instances can encourage the model to learn from low-PVI instances that require more complex inference, thereby enhancing generalization ability and reducing reliance on artifacts.
In the experiment, the Chinese-BERT-wwm model was used to calculate the PVI of the training set, and high-PVI instances were reduced in descending order of PVI by 10%, 20%, …, 90%, respectively, to construct training subsets with 90%, 80%, …, 10% of the original size. A series of experiments were conducted, we focused on analyzing the accuracy changes in the classification model at different reduction ratios in Table 2, and Table 3, Table 4 and Table 5 record the accuracy results on different datasets with different models, where SIM represents the standard input model, EIM represents the empty input model, and CM represents the classification model. As the reduction ratio of high-PVI instances increases, the accuracy of the classification models on the three datasets generally shows a declining trend, but the rate and extent of the decline vary across datasets, revealing the moderating effect of different task types on data redundancy. Figure 3 shows this trend.
OCNLI: As the easy instances are reduced, the accuracy of the model on the test set gradually decreases from 69.59% of the full training set to 22.97% (marked in red font in Table 3), with performance loss increasing linearly with the proportion of training set reduction. When removing 10–20% of high-PVI instances, the accuracy of the model decreases slightly (69.59%→68.85%→66.2%), indicating limited dependence of model performance on a small number of high-PVI instances. At this stage, the reduced dataset can save training resources while maintaining model performance within an acceptable range. After reducing 10% of the data, training time decreases, but accuracy drops by only 0.74%, meeting the practical application requirements for balancing efficiency and effectiveness. When 50% of the high-PVI instances are removed, the model accuracy drops to 49.37% (marked in blue font in Table 3), representing a decrease of 19.48% compared to removing 10% of the instances. This indicates that high-PVI instances still contain key generalizable information for the task, and excessive removal can disrupt the model’s ability to learn fundamental semantic patterns. The reason might be that not all high-PVI instances correspond to artifacts; some high PVI may arise from genuine strong correlations between input and labels (e.g., the logical relationship of “raining→wet ground” with “entailment” labels), and removing these instances would lead to information loss. Additionally, low-PVI instances contain complex inference patterns but may also include labeling noise or semantic ambiguity. Excessive removal of high-PVI instances alters the data distribution, directly increasing task difficulty beyond the model’s processing capacity, resulting in performance collapse.
Table 3. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in OCNLI.
Table 3. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in OCNLI.
OCNLIBase0.10.20.30.40.50.60.70.80.9
SIM *89.2983.4583.0878.2066.5664.0457.7057.5263.6872.11
EIM *34.1237.0942.5746.8545.9445.5331.9844.9545.2745.94
CM *69.5968.8566.2062.6054.2449.3741.2834.4226.8022.97
* SIM: standard input model. EIM: empty input model. CM: classification model.
Experiments demonstrate that high-PVI instances are irreplaceable for training the OCNLI model when the reduction ratio r ≥ 0.1, for the following reasons:
  • Loss of fundamental features: High-PVI instances typically contain strong association patterns between labels and inputs (e.g., the mapping of negation words like “不 (No)” to contradiction-class labels), which serve as the foundation for the model to learn basic inference rules. Removing these patterns makes it difficult for the model to learn basic inference rules.
  • Increased exposure to noise: Potential labeling errors or semantic ambiguity in low-PVI instances (e.g., ambiguous instances labeled as “neutral”) are amplified during training, disrupting the model’s optimization direction [38]. The removal of high-PVI instances disrupts the stable state of the original data distribution, where the noise dominates the training data, leading the model to converge to local optima. This result validates the core tenet of V -information theory: the difficulty of a dataset is a dynamic function of model capability and data distribution. The removal of high-PVI instances alters the data distribution, thereby changing the task difficulty.
OCNLI is a low-structured task, necessitating the retention of more high-PVI instances to maintain basic inference capabilities. When reducing the data, attention must be paid to the safe reduction ratio r of low-proportion deletion. Removing 10–20% of high-PVI instances results in only a slight decrease in accuracy on the test set (2–3% drop), making a reduction ratio of around 10% more recommended. The removal of a small number of high-PVI instances can eliminate some redundant artifacts (e.g., overly obvious syntactic templates), prompting the model to learn more generalizable features. However, the reduction ratio must be strictly limited (<20%), and a conservative reduction strategy should be adopted. Beyond 20%, the combined effect of fundamental feature loss and increased noise exposure would accelerate performance decline.
CMNLI: Without considering the balance of the dataset, as the reduction ratio increases, the accuracy of the model on the test set gradually decreases, from 79.99% when using the complete training set to 17.27% after removing 90% of easy instances (marked in red font in Table 4), which indicates that many easy instances being removed negatively impacts model performance. Among these, when 10% of high-PVI instances are removed, the accuracy is 79.94%, when 20% are removed, it is 79.23%, and when 30% are removed, it is 79.03% (marked in blue font in Table 4). This is similar to the experimental results on the OCNLI dataset, suggesting that the model’s performance has limited dependence on a small number of high-PVI instances. At this point, trimming the dataset can save training resources to some extent while maintaining model performance within an acceptable range. However, when more than 50% of the high-PVI instances are removed, the accuracy drops significantly, such that when 50% are removed, the accuracy is 0.6199 (marked in green font in Table 4), which is 17.95% lower than when 10% are removed. This may be because excessive removal leads to the loss of basic features, making it difficult for the model to effectively learn the semantic patterns, and the noise in low-PVI instances is amplified, affecting the model’s optimization direction.
Table 4. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in CMNLI.
Table 4. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in CMNLI.
CMNLIBase0.10.20.30.40.50.60.70.80.9
SIM88.5887.0684.6682.0174.5252.7548.0040.7451.6564.70
EIM33.3436.3636.4336.2236.9337.9738.9039.9340.7740.71
CM79.9979.9479.2379.0376.9461.9934.9437.3023.2917.27
CINLI: Through the use of the same static reduction method, the top 10%, 20%, …, 90% of high-PVI instances were removed in descending order of PVI to construct training subsets. The experimental results show that even after removing 40% of the high-PVI instances, the model’s accuracy remained at a high level of 87.13% (marked in red font in Table 5). This phenomenon contrasts significantly with the OCNLI experimental results, indicating a much slower performance degradation compared to OCNLI, revealing the regulatory effect of task types on data redundancy.
Table 5. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in CINLI.
Table 5. Accuracy (%) comparison between different reduction ratios ( r from 0 to 0.9) in CINLI.
CINLIBase0.10.20.30.40.50.60.70.80.9
SIM97.3297.0396.3195.2892.3288.1786.0780.0759.9359.93
EIM29.0737.6142.3247.9355.9264.8256.4746.0428.5336.34
CM91.1491.3190.7689.0487.1377.7079.0575.5348.7048.70
The stability of CINLI stems from its intrinsic characteristics:
  • Structured semantics: The fixed meaning of idioms allows the model to perform generalization inference using a small number of keywords (e.g., “剑 (sword)” in “刻舟求剑”, which literally means “to carve a mark on a boat to find a lost sword”; or “蛇 (snake)” and “足 (foot)” in “画蛇添足”, which means “to draw a snake and add feet to it”), reducing reliance on data volume and eliminating the need to learn complex contextual correlations. This differs from the causal chain inference in OCNLI, where complex logical inference also requires more task-specific parameter updates. Additionally, the semantic boundaries of idioms are clear, resulting in higher compactness of data distribution and a more concentrated PVI distribution of training instances (low redundancy in high-PVI instances). Even after removal, the remaining instances still cover core semantic patterns.
  • Pre-training compensation: BERT-wwm has encoded the general semantics of idioms [28], thereby reducing sensitivity to training instances. The idiom inference task in CINLI is highly compatible with BERT’s masked language modeling objective, both relying on local semantic correlations. Through large-scale corpora, idioms have learned distributed representations, and model finetuning only requires aligning the label space rather than constructing semantic mappings from scratch. Therefore, even after removing some instances, the model can still leverage prior knowledge for generalization inference. This phenomenon aligns with the discussion in the original text on the task–distribution coupling effect: task difficulty is determined by both data distribution attributes (e.g., degree of semantic structuring) and model prior knowledge.
CINLI corresponds to highly structured tasks, with strong feasibility of data reduction, allowing for the prioritized removal of redundant high-PVI instances, saving resources without affecting performance. For such tasks, an aggressive reduction strategy can be adopted, which can reduce approximately 30% of high-PVI instances.
Class Balance: In the process of reducing the dataset, we discovered that as more easy instances were reduced, more class imbalances were introduced in the remaining training subset. Therefore, we artificially controlled for proportional reduction in each category and explored the impact of class balance on model training. The experimental results (see Appendix A) indicate that after applying balanced reduction to the dataset to balance the class distribution, the issue of distribution bias caused by the removal of high-PVI instances was mitigated to some extent. Under this balanced constraint, the accuracy of the trained empty input model (EIM) consistently remained close to the random probability of a three-class classification (33%), which aligns with our assumption about balanced reduction. This suggests that the balanced constraint effectively weakens the impact of label distribution bias but does not alter the information-theoretic nature of the empty model. The limited utilization of input information by the empty model and the stability of its performance further highlight the capability of standard input models in effectively utilizing input information for prediction. Simultaneously, this also indirectly confirms that the performance decline of the standard input model after the removal of high-PVI instances is not due to the model itself becoming completely ineffective, but rather because it loses the effective utilization of key input information.
Noise: To verify the robustness and generalization ability of the method, we introduced noise into the OCNLI dataset, aiming to simulate more realistic application scenarios. We randomly replaced instances in the OCNLI training set with low-quality text at a replacement ratio of 0.1. These low-quality texts were generated by rewriting original sentences, incorporating features such as synonym substitution, misspellings, punctuation noise, internet slang, meaningless phrase insertion, and sentence splitting and merging. As depicted in Figure 4, the model’s performance trend on the noisy dataset closely mirrors that on both class-imbalanced and class-balanced datasets. According to the previous conclusion, a reduction of at least 10% of the data, as per PVI, has little impact on the model’s performance. It is evident that the performance on datasets with added noise almost remains lower than that on the other two dataset types (imbalanced and balanced).
Random: We also included a random baseline as a control, randomly deleting instances from the training data without considering any scores or specific features of the data points. We observed that reducing the easy instances did not lead to a gain in performance over the random baseline. A similar phenomenon appeared in Rabiraj’s research [39], which designed a pruning strategy for sexism detection using three influence scores including PVI. As the reduction ratio of simple instances increased, the training subset contained more difficult instances. Consequently, the difficulty of the dataset rose, and the performance gap between the random baseline and other baselines using PVI for reduction gradually widened. We speculated that training solely on a difficult subset could cause the model to over-emphasize learning from edge cases and ambiguous instances, potentially leading to overfitting to these specific hard instances and consequently poor generalization on the broader test set. The random baseline’s performance, where some easy instances were inevitably retained due to random deletion, might implicitly benefit from this broader representation. Removing a large number of simple instances could disrupt the difficulty distribution of the dataset. Therefore, we recommend reducing the dataset within a moderate range, removing simpler instances only to the extent that the overall difficulty structure remains largely intact. To further improve upon the baseline performance and the reduction in simple instances, one could incorporate a more sophisticated weighting mechanism for the remaining difficult instances during training. In subsequent work, we will consider combining PVI-based difficulty measurement with diversity measurement (e.g., EL2N [40], VoG [41], TracIn [42]) as an improvement strategy to select the instances to be retained. This would ensure that the model learns not only from challenging instances but also from a diverse set of representative instances across the data spectrum.

3.2.2. Progressive Learning

In this section, the experiments primarily focus on the OCNLI and CINLI datasets, aiming to investigate the effectiveness of progressive learning strategies. The selection of these two datasets is based on the following considerations: The OCNLI dataset holds significant representativeness in the field of Chinese natural language inference, effectively evaluating the model’s baseline performance and generalization capabilities; the CINLI dataset, with its unique text pair construction and inference task design, facilitates an in-depth examination of the model’s inference accuracy and stability. In comparison, the CMNLI dataset, with its large instance size and status as a translation-generated dataset, exhibits limitations such as semantic bias and cultural differences, which may introduce confounding factors. Therefore, under constrained experimental resources, prioritizing the OCNLI and CINLI datasets ensures the acquisition of more reference-worthy and persuasive experimental results.
Following Algorithm 3, the training set is sorted based on PVI (from easiest to hardest), and Qwen3-0.6B (available on https://huggingface.co/Qwen/Qwen3-0.6B (accessed on 15 June 2025)) is used as the base model to train. Initially, PVI values are computed for all instances in the training set to establish their difficulty ranking. Then, the training process commences with the simplest instances and gradually incorporates more difficult ones by selecting subsets of the sorted training data. After each progressive training stage on a subset, the trained model is evaluated on a fixed held-out test set, recording accuracy, precision, recall, and F1 score to assess performance evolution. The experimental results demonstrate that training the dataset sorted by PVI enhances model performance. Since Micro-average is used to calculate recall in multi-class tasks, the three categories in the dataset are relatively evenly distributed, with values close to accuracy.
Table 6 presents the experimental results on the OCNLI dataset. By sorting the training set based on PVI from easiest to hardest, the model’s accuracy improves by approximately 0.81% relative to the baseline (69.76% − 68.95 = 0.81%), and the F1 score also rises from a baseline of 69.08% to 69.91%, an increase of about 0.83% (marked in bold font in Table 6). This indicates a positive impact of PVI sorting on model performance. Even with a 10% reduction in training data, the model performance remains high, reflecting the effectiveness of the sorting and reduction strategies. Table 7 presents the mean and standard deviation of the model’s performance over three runs on the OCNLI dataset. Figure 5 visually compares the model’s performance under different processing methods. Comparing the “Base” (green bar) and “Sort” (blue bar) clearly shows that after PVI sorting, the model improves in accuracy, precision, recall, and F1 score. While “Sort & Reducing 10%” (orange bar) performs slightly lower than “Sort” on all metrics, it still maintains a level of precision close to that of “Base,” consistent with the data analysis in Table 6, further confirming that even with reduced data volume, the model can still exhibit strong performance.
Table 8 presents the experimental results on the CINLI dataset, showing that the model’s performance also improves after data processing. The accuracy increases from the baseline of 91.7852% to 91.8676%. The F1 score rises from the baseline of 91.7861% to 91.8651% (marked in bold font in Table 8). At a reduction ratio of r =0.3 (i.e., reducing 30% of the data volume), the model still maintains an accuracy of 90.42% and an F1 score of 90.38%, further validating that the progressive learning strategy can effectively reduce the demand for training data while preserving model performance. Figure 6 compares the model’s performance on the CINLI dataset under different processing methods. Similarly to the analysis of OCNLI, Figure 6 clearly illustrates the improvements in “Sort” (blue bar) over “Base” (green bar) in all performance metrics, although the magnitude of the improvement is relatively small. It is noteworthy that the performance of “Sort & Reducing 30%” (orange bar) declines in accuracy, precision, recall, and F1 score, but remains above 90%.
Table 9 presents the mean and standard deviation of the model’s performance over three runs on the CINLI dataset.
We speculate that this progressive learning strategy from easy to difficult (as a form of curriculum learning) enables the model to prioritize learning instances that are information-rich but low in difficulty during the early stages of training, thereby rapidly constructing foundational feature representations and pattern recognition capabilities. Subsequently, the model gradually exposes itself to and learns more complex instances, which helps it progressively master more abstract and fine-grained knowledge. This reasonable distribution of difficulty optimizes the “quality” and utilization efficiency of the training set during the training process, avoiding interference from a large number of difficult or noisy instances in the early stages, thus promoting faster convergence rates and higher final performance. From the perspective of model optimization, a reasonable distribution of difficulty can guide the gradient descent process to converge to better local minima or, at the very least, achieve more robust parameter initialization in the early stages of training, laying a solid foundation for subsequent learning.

3.3. Computational Efficiency Analysis

Figure 7 illustrates the empty input model’s computation time for each dataset at various reduction ratios, using a single NVIDIA A100 GPU (NVIDIA, Santa Clara, CA, USA). The EIM’s runtime decreases as the dataset size is reduced. While data reduction decreases the training time, the initial PVI computation requires some overhead, particularly at lower reduction rates. Where there is an initial cost, we should consider the breakeven point and subsequent benefits of training, especially when considering the possibility of larger reductions or multiple training runs on the same reduced dataset. This overhead cost may be acceptable, depending on the overall efficiency gains.

4. Discussion

This chapter discusses the reasons why the NLI dataset poses challenges for model construction, which may stem from the inherent characteristics of the dataset and its intrinsic distribution.
In the CMNLI dataset, the distribution of token counts across different inference categories is unbalanced. Figure 8 shows the histogram of hypothesis length distribution in the CMNLI dataset, and combined with the statistical information in Table 10, it can be observed that the statistical features of neutral hypotheses (label 1) are significantly different from other categories, with a median (17.0) and mean (18.35) that are both the highest, and a maximum value reaching 100 tokens. This indicates that the length distribution of neutral hypotheses is right-skewed, reflecting that maintaining semantic neutral status requires more modifiers, such as adding conditional adverbials (“under certain conditions”) or hedges (“possibly”), leading to neutral hypotheses containing the longest instances. Therefore, hypothesis length becomes an effective feature, with neutral hypotheses dominating the longer text intervals (e.g., constituting 27.20% in the 16–20-token range, and consistently leading in intervals ≥31 tokens). Contradiction hypotheses (label 2) exhibit the shortest concentration trend, with a median of 15.0. The syntactic characteristic of Chinese, known as “parataxis,” allows for the expression of complex logic using fewer tokens, which may have influenced the conciseness of contradiction hypothesis categories.
The proportions of assumptions in different length intervals of the CMNLI dataset are presented in Table 11. In the short text interval (1–10 tokens), the proportions of entailment (0.201) and contradiction (0.225) are slightly higher than neutral (0.129). Short texts do not exhibit a clear category advantage, which is related to the characteristics of the Chinese language. In the core distribution interval (11–25 tokens), the neutral hypothesis has the highest proportion (27.20%) in the 16–20-token range, while the contradiction hypothesis forms a peak in the 11–15-token range (30.9%), with this region showing cross-competition among the three types of hypotheses. In the ultra-long text interval (≥31 tokens), the neutral hypothesis maintains a leading proportion, significantly higher than entailment and contradiction.
For OCNLI, as observed from the distribution histogram in Figure 9 and sentence length statistics in Table 12, the distribution of hypothesis lengths in the OCNLI dataset exhibits a clear right skew, with most hypotheses concentrated in shorter length intervals (particularly 5–15 tokens). Compared to the CMNLI dataset, OCNLI’s hypotheses are generally shorter, and even the longest instances are significantly shorter (maximum value of 60 tokens). The construction of the OCNLI dataset emphasizes shorter, more direct inference scenarios, and the characteristics of its text sources also contribute to the shorter hypotheses.
The proportion of different length intervals of assumed content in the OCNLI dataset is presented in Table 13. In the short text interval (1–5 tokens), the proportion of entailment assumptions (0.083) and contradiction assumptions (0.062) is significantly higher than that of neutral assumptions (0.046). This indicates that in the OCNLI dataset, short texts seem to better support entailment and contradiction relationships, which may be related to certain phrases or expressions in Chinese that can directly constitute entailment or contradiction relationships. The core distribution interval (6–15 tokens) is a very concentrated interval, with all three types of assumptions accounting for most instances. The 6–10-token interval is the peak: entailment (0.479), neutral (0.425), and contradiction (0.475) all reach their respective peaks in this interval, with proportions all close to or exceeding 40%. This suggests that the core assumption length in the OCNLI dataset is concentrated between 6 and 10 tokens. In the 11–15-token interval, the proportion of neutral assumptions (0.337) is the highest, slightly exceeding that of contradiction (0.321) and entailment (0.304). This again confirms the trend that neutral assumptions tend to be relatively longer. In the medium–long text interval and the ultra-long text interval, the advantage of neutral assumptions gradually becomes apparent, with proportions consistently leading those of entailment and contradiction assumptions. These results further demonstrate that maintaining neutral status requires longer expressions or neutral inference in more complex contexts.
Furthermore, we have listed the most challenging instances from the OCNLI test set according to Chinese-BERT-wwm, detailed in Appendix B. All three categories—entailment, neutral, and contradiction—are represented in Table A2, with entailment representation appearing slightly excessive. Some instances have actually been incorrectly labeled—for instance, the instance “Premise: 他是去那个南方那个学校嘛 (He is going to that southern school, right?) Hypothesis: 国防动员无需加强 (National defense mobilization does not need to be strengthened)” is labeled as “entailment,” although the correct label should be “neutral.”

5. Conclusions and Future Work

We introduced an effective data reduction strategy based on Pointwise V-Information (PVI) to enhance model training efficiency and performance in data-centric AI. We successfully extended the PVI framework, previously limited to English datasets, to various Chinese NLP tasks and base models, addressing a critical gap in cross-lingual data reduction.
The use of PVI in this article also has certain limitations. A key challenge is that PVI requires the model to produce a full probability distribution over all possible outputs. This can be problematic for tasks where such distributions are not readily available, such as machine translation that relies on beam search. In addition, while PVI enables comparison of different instances with respect to the same attribute (e.g., the amount of token-identity information available for two instances), it does not support direct comparison of different attributes for the same instance.
In future work, we aim to validate the generality of our method in diverse linguistic environments by conducting experiments on German, French, and other multilingual NLI corpora. To this end, we will explore and utilize pre-trained language models suitable for these languages, such as GottBERT (for German) and FlauBERT (for French). This comprehensive evaluation will allow us to gain a deep understanding of the PVI-based strategy’s capabilities for data reduction and performance enhancement in cross-lingual scenarios, as well as identify potential challenges across different language characteristics and task contexts.
We acknowledge that the optimal data reduction approach may vary significantly across different data modalities, such as text, images, or tabular data. Therefore, tailoring reduction methods to the unique characteristics of different data types and application domains will be crucial. The information-theoretic basis of PVI indicates its relevance beyond textual data. We anticipate expanding PVI’s application to non-text modalities, including visuals and speech. Applying PVI to picture data necessitates delineating the extraction of model-usable information from pixel arrays and the construction of “empty inputs” for visual tasks. Likewise, with audio data, PVI might potentially utilize spectral information or alternative acoustic representations. Extending PVI to these modalities necessitates meticulous evaluation of modality-specific attributes, including the establishment of suitable null inputs or baselines for computing V-information that accurately represent the computational limitations of models functioning with these data types. Comprehending these modality-specific adaptations and constraints will be essential for our future study, aiding in the evaluation of PVI’s generalizability as a universal standard for data quality.
Data selection encompasses multiple criteria, such as entropy-based techniques, loss-based data trimming, gradient-based sampling, and methods inspired by active learning, among others. We also plan to move beyond single indicators for data point selection, aiming for more nuanced metrics that capture a data point’s value in terms of diversity, informativeness, or representativeness of important subgroups.
Additionally, an exciting direction involves combining data reduction with synthetic data generation. The Phi-series models [43,44] from Microsoft demonstrate that carefully selected and optimized training datasets can substantially decrease model size without compromising performance. Synthetic data, often produced by advanced LLMs, is utilized to augment these datasets, such as by closing performance gaps or delivering specialized skill training. Future systems could identify gaps created by aggressive filtering, especially for rare but important instances, and then use generative models to create synthetic data to fill these specific gaps, ensuring comprehensive coverage.

Author Contributions

Conceptualization, F.C. and W.Z.; methodology, F.C. and W.Z.; software, W.Z.; validation, F.C. and W.Z.; formal analysis, F.C.; investigation, F.C.; resources, W.Z.; data curation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, F.C.; visualization, W.Z.; supervision, F.C.; project administration, F.C.; funding acquisition, F.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the basic scientific research project of the Educational Department of Liaoning Province, the People’s Republic of China. The grant number is 984230020.

Data Availability Statement

Data are shown within the text and in the appendix.

Acknowledgments

The authors thank the Information Science and Technology College at Dalian Maritime University for supporting this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Impact of label distribution bias on the model. The accuracy of the trained empty input model (EIM, marked in bold font in Table A1) consistently remained close to the random probability of a three-class classification (33%).
Table A1. Impact of label distribution bias on the model. The accuracy of the trained empty input model (EIM, marked in bold font in Table A1) consistently remained close to the random probability of a three-class classification (33%).
DatasetModelBase0.10.20.30.40.50.60.70.80.9
OCNLISIM89.2986.6580.3679.7471.4557.9234.1759.9463.8070.65
EIM34.1233.7933.4033.1833.0132.5732.9834.1733.6334.43
CM69.5968.5667.2267.7364.4757.7434.1635.1127.7523.82
CMNLISIM88.5887.4284.2381.2174.0742.2533.3234.4533.3461.05
EIM33.3433.3433.3433.3433.3432.9733.3233.7633.2933.32
CM79.9979.9378.9278.7876.6047.4132.0733.6834.9419.47
CINLISIM97.3296.9297.3294.4192.5889.0379.6033.5644.1392.58
EIM29.0733.5627.0929.6533.8633.8633.5633.5632.1033.86
CM91.1491.1391.1490.4087.2683.7375.2233.7841.0887.26

Appendix B

Certain instances within Table A2 are assessed to be mislabeled by the authors of this work, and these are visually indicated in red. The labels of the fifth and ninth instances are entailment, which should be changed to neutral. The label of the 11th instance is neutral, which should be changed to entailment.
Table A2. Part of the hardest (lowest-PVI) instances in the OCNLI test set for logical inference (label indicates the logical relationship between “premise” and “hypothesis”), according to Chinese-BERT-wwm. Instances in red are assessed to be mislabeled by the authors of this work.
Table A2. Part of the hardest (lowest-PVI) instances in the OCNLI test set for logical inference (label indicates the logical relationship between “premise” and “hypothesis”), according to Chinese-BERT-wwm. Instances in red are assessed to be mislabeled by the authors of this work.
NumPremiseHypothesisLabelPVI
1其中有一个这两天记者采访他还出诊呢记者工作是出诊矛盾−8.745
One of them was interviewed by a journalist and even made house calls these last two daysJournalists make house callsContradiction
2所以对热闹的世界杯充耳不闻“我们”没有关注世界杯蕴含−7.125
So I turned a deaf ear to the lively World Cup“We" did not pay attention to the World CupEntailment
3处理中美关系应着眼于全球,着眼于二十一世纪二十世纪对处理中美关系不重要中立−6.645
Handling China-US relations should be focused on the global perspective, focused on the 21st centuryThe 20th century is not important for handling China-US relationsNeutral
4然而,古巴、也门和其它一些国家一直要求立即取消对伊拉克的制裁其他一些国家包括古巴和也门矛盾−6.622
However, Cuba, Yemen, and some other countries have consistently demanded the immediate lifting of sanctions on IraqOther countries include Cuba and YemenContradiction
5他是去那个南方那个学校嘛国防动员无需加强蕴含−6.561
Is he going to that school in the south National defense mobilization does not need to be strengthenedEntailment
6对外承包工程和劳务合作完成营业额近13亿美元营业额达到了13亿美元矛盾−6.545
Contracted engineering projects and labor cooperation completed nearly $1.3 billion in turnoverRevenue reached $1.3 billionContradiction
7去年8月海湾冲突爆发后,日本政府曾向国会提出了一项旨在向海外派兵的联合国和平合作法案日本政府没有独立的立法权蕴含−6.333
Last August after the Gulf conflict broke out, the Japanese government proposed a UN Peace Cooperation Bill to the Diet aimed at deploying troops overseasThe Japanese government does not have independent legislative powerEntailment
8各项决策都要做到程序依法规范、过程民主公开、结果科学公正没有一条好决策的出台能够脱离依法规范的程序蕴含−6.291
All decisions must be made in accordance with legally standardized procedures, democratic and open processes, and scientifically fair outcomesNo good decision can be implemented without reference to legally standardized proceduresEntailment
9花篮里的花又白的多红的少,专配银冠似的我对花盆的花颜色的搭配嗤之以鼻蕴含−6.260
The flowers in the basket are mostly white and few red, perfectly matching the silver crown-like appearanceI sneer at the color combination of the flowers in the potsEntailment
10合理的投资规模是保持经济稳定和增强发展后劲的重要条件不合理的投资规模制约经济持续向好蕴含−5.989
A reasonable investment scale is an important condition for maintaining economic stability and enhancing development potentialAn unreasonable investment scale restricts the economy’s sustainable upward trendEntailment
11对,出席大会的时候还自我调侃,说这个整个场面,我是个科学家,我不是摇滚明星举办过一场大会中立−5.944
Yes, when attending the conference, I even made a self-deprecating joke, saying that in this entire scene, I’m a scientist, not a rock starOrganized a conferenceNeutral

Appendix C

A comprehensive list and description of all symbols employed in this paper can be found in Table A3.
Table A3. List of notations and their descriptions.
Table A3. List of notations and their descriptions.
NotationsDescriptions
X Sentence
X Random variables taking values in X
Y Label
Y Random variables taking values in Y
( x , y ) An instance in the dataset
V Predictive family
f ( ) Predictive function in V
g The best predictive function for a given input X
g The best predictive function for output Y without specific input X
Empty input
H V ( ) V -entropy
D Complete dataset
D e a s y Dataset composed of simple instances
D e a s y A subset selected from D e a s y
J ( θ ) Loss function
τ Threshold for defining simple instances
ϵ Minimum value of performance gap
I ( ) Shannon’s mutual information
r Reduction ratio
m Total number of instances in the training dataset
P ( ) The set of all probability measures over the Borel algebra
t X X , being any function

References

  1. Zhou, Y.; Tu, F.; Sha, K.; Ding, J.; Chen, H. A Survey on Data Quality Dimensions and Tools for Machine Learning. arXiv 2024. [Google Scholar] [CrossRef]
  2. Smart Trends in Computing and Communications. In Proceedings of the SmartCom 2024, Pune, India, 12–13 January 2024; Volume 1. SpringerLink. Available online: https://link.springer.com/book/10.1007/978-981-97-1320-2 (accessed on 15 June 2025).
  3. Wang, Z.; Shang, J.; Liu, L.; Lu, L.; Liu, J.; Han, J. CrossWeigh: Training Named Entity Tagger from Imperfect Annotations. In Proceedings of the EMNLP-IJCNLP 2019, Hong Kong, China, 3–7 November 2019. [Google Scholar] [CrossRef]
  4. Northcutt, C.; Jiang, L.; Chuang, I. Confident Learning: Estimating Uncertainty in Dataset Labels. J. Artif. Intell. Res. 2021, 70, 1373–1411. [Google Scholar] [CrossRef]
  5. Song, H.; Kim, M.; Park, D.; Shin, Y.; Lee, J.G. Learning From Noisy Labels With Deep Neural Networks: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 8135–8153. [Google Scholar] [CrossRef] [PubMed]
  6. Gudivada, V.N.; Apon, A.; Ding, J. Data Quality Considerations for Big Data and Machine Learning: Going Beyond Data Cleaning and Transformations. Int. J. Adv. Softw. 2017, 10, 1–20. [Google Scholar]
  7. Joshi, S.; Jain, A.; Payani, A.; Mirzasoleiman, B. Data-Efficient Contrastive Language-Image Pretraining: Prioritizing Data Quality over Quantity. In Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 18 April 2024; pp. 1000–1008. [Google Scholar] [CrossRef]
  8. Bhatt, N.; Bhatt, N.; Prajapati, P.; Sorathiya, V.; Alshathri, S.; El-Shafai, W. A Data-Centric Approach to Improve Performance of Deep Learning Models. Sci. Rep. 2024, 14, 22329. [Google Scholar] [CrossRef]
  9. Toscano-Durán, V.; Perera-Lago, J.; Paluzo-Hidalgo, E.; Gonzalez-Diaz, R.; Gutierrez-Naranjo, M.Á.; Rucco, M. An In-Depth Analysis of Data Reduction Methods for Sustainable Deep Learning. Open Res. Eur. 2024, 4, 101. [Google Scholar] [CrossRef] [PubMed]
  10. Lei, S.; Tao, D. A Comprehensive Survey of Dataset Distillation. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 17–32. [Google Scholar] [CrossRef]
  11. Hoffmann, J.; Borgeaud, S.; Mensch, A.; Buchatskaya, E.; Cai, T.; Rutherford, E.; Casas, D.d.L.; Hendricks, L.A.; Welbl, J.; Clark, A.; et al. Training Compute-Optimal Large Language Models. In Proceedings of the 36th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 28 November 2022. [Google Scholar] [CrossRef]
  12. Kwok, D.; Anand, N.; Frankle, J.; Dziugaite, G.K.; Rolnick, D. Dataset Difficulty and the Role of Inductive Bias. arXiv 2024. [Google Scholar] [CrossRef]
  13. Baldock, R.; Maennel, H.; Neyshabur, B. Deep Learning Through the Lens of Example Difficulty. Adv. Neural Inf. Process. Syst. 2021, 34, 10876–10889. [Google Scholar]
  14. Hooker, S.; Moorosi, N.; Clark, G.; Bengio, S.; Denton, E. Characterising Bias in Compressed Models. arXiv 2020. [Google Scholar] [CrossRef]
  15. Cui, P.; Zhang, D.; Deng, Z.; Dong, Y.; Zhu, J. Learning Sample Difficulty from Pre-Trained Models for Reliable Prediction. Adv. Neural Inf. Process. Syst. 2023, 36, 25390–25408. [Google Scholar]
  16. Mayo, D.; Cummings, J.; Lin, X.; Gutfreund, D.; Katz, B.; Barbu, A. How Hard Are Computer Vision Datasets? Calibrating Dataset Difficulty to Viewing Time. Adv. Neural Inf. Process. Syst. 2023, 36, 11008–11036. [Google Scholar]
  17. Wang, C.; Dong, Q.; Wang, X.; Wang, H.; Sui, Z. Statistical Dataset Evaluation: Reliability, Difficulty, and Validity. arXiv 2022. [Google Scholar] [CrossRef]
  18. Ethayarajh, K.; Choi, Y.; Swayamdipta, S. Understanding Dataset Difficulty with V-Usable Information. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; Volume 162, pp. 5988–6008. [Google Scholar]
  19. Havrilla, A.; Dai, A.; O’Mahony, L.; Oostermeijer, K.; Zisler, V.; Albalak, A.; Milo, F.; Raparthy, S.C.; Gandhi, K.; Abbasi, B.; et al. Surveying the Effects of Quality, Diversity, and Complexity in Synthetic Data From Large Language Models. arXiv 2024. [Google Scholar] [CrossRef]
  20. Recht, B.; Roelofs, R.; Schmidt, L.; Shankar, V. Do ImageNet Classifiers Generalize to ImageNet? In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 24 May 2019; pp. 5389–5400. Available online: https://proceedings.mlr.press/v97/recht19a.html (accessed on 15 June 2025).
  21. Hendrycks, D.; Zhao, K.; Basart, S.; Steinhardt, J.; Song, D. Natural Adversarial Examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 11–15 June 2021; pp. 15262–15271. [Google Scholar] [CrossRef]
  22. Levow, G.A. The Third International Chinese Language Processing Bakeoff: Word Segmentation and Named Entity Recognition. In Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing, Sydney, Australia, 22–23 July 2006; pp. 108–117. [Google Scholar]
  23. Xu, L.; Tong, Y.; Dong, Q.; Liao, Y.; Yu, C.; Tian, Y.; Liu, W.; Li, L.; Liu, C.; Zhang, X. CLUENER2020: Fine-Grained Named Entity Recognition Dataset and Benchmark for Chinese. arXiv 2020. [Google Scholar] [CrossRef]
  24. Yu, F.; Zhang, H.; Tiwari, P.; Wang, B. Natural Language Reasoning, A Survey. ACM Comput. Surv. 2024, 56, 1–39. [Google Scholar] [CrossRef]
  25. Xu, Y.; Zhao, S.; Song, J.; Stewart, R.; Ermon, S. A Theory of Usable Information Under Computational Constraints. International Conference on Learning Representations. Available online: https://openreview.net/forum?id=r1eBeyHFDH (accessed on 15 June 2025).
  26. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  27. Cui, Y.; Che, W.; Liu, T.; Qin, B.; Yang, Z. Pre-Training With Whole Word Masking for Chinese BERT. IEEE/ACM Trans. Audio Speech Lang. Process 2021, 29, 3504–3514. [Google Scholar] [CrossRef]
  28. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 3–5 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  29. Cui, Y.; Che, W.; Liu, T.; Qin, B.; Wang, S.; Hu, G. Revisiting Pre-Trained Models for Chinese Natural Language Processing. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16–20 November 2020; pp. 657–668. [Google Scholar] [CrossRef]
  30. Fayek, H.M.; Cavedon, L.; Wu, H.R. Progressive Learning: A Deep Learning Framework for Continual Learning. Neural Netw. 2020, 128, 345–357. [Google Scholar] [CrossRef]
  31. Mao, A.; Mohri, M.; Zhong, Y. Cross-Entropy Loss Functions: Theoretical Analysis and Applications. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 23803–23828. Available online: https://proceedings.mlr.press/v202/mao23b.html (accessed on 15 June 2025).
  32. Nguyen, T. Understanding Transformers via N-Gram Statistics. Adv. Neural Inf. Process Syst. 2024, 37, 98049–98082. [Google Scholar]
  33. Yang, A.; Li, A.; Yang, B.; Zhang, B.; Hui, B.; Zheng, B.; Yu, B.; Gao, C.; Huang, C.; Lv, C.; et al. Qwen3 Technical Report. arXiv 2025. [Google Scholar] [CrossRef]
  34. Hu, H.; Richardson, K.; Xu, L.; Li, L.; Kübler, S.; Moss, L.S. OCNLI: Original Chinese Natural Language Inference. arXiv 2020. [Google Scholar] [CrossRef]
  35. Xu, L.; Hu, H.; Zhang, X.; Li, L.; Cao, C.; Li, Y.; Xu, Y.; Sun, K.; Yu, D.; Yu, C.; et al. CLUE: A Chinese Language Understanding Evaluation Benchmark. In Proceedings of the 28th International Conference on Computational Linguistic, Barcelona, Spain, 8–13 December 2020; pp. 4762–4772. [Google Scholar] [CrossRef]
  36. Conneau, A.; Rinott, R.; Lample, G.; Williams, A.; Bowman, S.; Schwenk, H.; Stoyanov, V. XNLI: Evaluating Cross-Lingual Sentence Representations. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 2–4 November 2018; pp. 2475–2485. [Google Scholar] [CrossRef]
  37. Williams, A.; Nangia, N.; Bowman, S.R. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. arXiv 2018. [Google Scholar] [CrossRef]
  38. Gururangan, S.; Swayamdipta, S.; Levy, O.; Schwartz, R.; Bowman, S.; Smith, N.A. Annotation Artifacts in Natural Language Inference Data. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; Volume 2, pp. 107–112. [Google Scholar] [CrossRef]
  39. Bandyopadhyay, R.; Assenmacher, D.; Moral, J.M.A.; Wagner, C. Sexism Detection on a Data Diet. In Proceedings of the Companion Publication of the 16th ACM Web Science Conference, Stuttgart, Germany, 21–24 May 2024; pp. 94–102. [Google Scholar] [CrossRef]
  40. Paul, M.; Ganguli, S.; Dziugaite, G.K. Deep Learning on a Data Diet: Finding Important Examples Early in Training. Adv. Neural Inf. Process Syst. 2021, 34, 20596–20607. [Google Scholar] [CrossRef]
  41. Agarwal, C.; D’souza, D.; Hooker, S. Estimating Example Difficulty Using Variance of Gradients. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10358–10368. [Google Scholar] [CrossRef]
  42. Pruthi, G.; Liu, F.; Kale, S.; Sundararajan, M. Estimating Training Data Influence by Tracing Gradient Descent. Adv. Neural Inf. Process. Syst. 2020, 33, 19920–19930. [Google Scholar]
  43. Abdin, M.; Aneja, J.; Awadalla, H.; Awadallah, A.; Awan, A.A.; Bach, N.; Bahree, A.; Bakhtiari, A.; Bao, J.; Behl, H.; et al. Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone. arXiv 2024. [Google Scholar] [CrossRef]
  44. Microsoft; Abouelenin, A.; Ashfaq, A.; Atkinson, A.; Awadalla, H.; Bach, N.; Bao, J.; Benhaim, A.; Cai, M.; Chaudhary, V.; et al. Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs. arXiv 2025. [Google Scholar] [CrossRef]
Figure 1. The model architecture of the data reduction strategy.
Figure 1. The model architecture of the data reduction strategy.
Electronics 14 03092 g001
Figure 2. The distribution of instance difficulty (PVI) in the held-out sets for each dataset.
Figure 2. The distribution of instance difficulty (PVI) in the held-out sets for each dataset.
Electronics 14 03092 g002
Figure 3. Accuracy varies with the reduction ratio. The x-axis represents the reduction ratio (the proportion of data removed), ranging from 0 to 0.9 (or 0% to 90%). The y-axis shows the accuracy (%), indicating the model’s performance on the test set. The legend includes three datasets: OCNLI, CMNLI, and CINLI. The figure visually demonstrates that initial data reductions have limited impact on accuracy, especially in the early stages. Our finding implies that training a classifier on the subset with 10–30% of the original data removed could preserve most of the accuracy while significantly improving training efficiency when combined with an effective data reduction strategy.
Figure 3. Accuracy varies with the reduction ratio. The x-axis represents the reduction ratio (the proportion of data removed), ranging from 0 to 0.9 (or 0% to 90%). The y-axis shows the accuracy (%), indicating the model’s performance on the test set. The legend includes three datasets: OCNLI, CMNLI, and CINLI. The figure visually demonstrates that initial data reductions have limited impact on accuracy, especially in the early stages. Our finding implies that training a classifier on the subset with 10–30% of the original data removed could preserve most of the accuracy while significantly improving training efficiency when combined with an effective data reduction strategy.
Electronics 14 03092 g003
Figure 4. Accuracy (%) variation between different ratios in OCNLI.
Figure 4. Accuracy (%) variation between different ratios in OCNLI.
Electronics 14 03092 g004
Figure 5. Comparison of indicators on the OCNLI dataset.
Figure 5. Comparison of indicators on the OCNLI dataset.
Electronics 14 03092 g005
Figure 6. Comparison of indicators on the CINLI dataset.
Figure 6. Comparison of indicators on the CINLI dataset.
Electronics 14 03092 g006
Figure 7. Runtime analysis of empty input models during the training process. The x-axis represents the reduction ratio r , ranging from 0 to 0.9, indicating the proportion of high-PVI instances removed during data reduction. The y-axis shows the training time (in seconds), reflecting the computational resources required for training under each r . The yellow dot indicates the calculation time of running the empty model on the class-unbalanced dataset, and the green dot indicates the calculation time of running the empty model on the class-balanced dataset.
Figure 7. Runtime analysis of empty input models during the training process. The x-axis represents the reduction ratio r , ranging from 0 to 0.9, indicating the proportion of high-PVI instances removed during data reduction. The y-axis shows the training time (in seconds), reflecting the computational resources required for training under each r . The yellow dot indicates the calculation time of running the empty model on the class-unbalanced dataset, and the green dot indicates the calculation time of running the empty model on the class-balanced dataset.
Electronics 14 03092 g007
Figure 8. Display of hypothesis length distribution for CMNLI.
Figure 8. Display of hypothesis length distribution for CMNLI.
Electronics 14 03092 g008
Figure 9. Display of hypothesis length distribution for OCNLI.
Figure 9. Display of hypothesis length distribution for OCNLI.
Electronics 14 03092 g009
Table 1. Category statistics of dataset usage quantity.
Table 1. Category statistics of dataset usage quantity.
DatasetSetTotalEntailment 1NeutralContradiction
OCNLItraining40,34013,464 (33.4%)13,734 (34.0%)13,142 (32.6%)
testing10,0973315 (32.8%)3448 (34.1%)3334 (33.0%)
CMNLItraining391,783130,612 (33.3%)130,555 (33.3%)130,616 (33.3%)
testing12,2414277 (32.9%)3926 (32.0%)4038 (32.9%)
CINLI *training80,12426,112 (32.5%)26,886 (33.5%)27,126 (33.8%)
testing26,7088634 (32.3%)9022 (33.7%)9052 (33.8%)
* CINLI is an open-source dataset maintained by individuals, which can be accessed through the GitHub repository (https://github.com/liucongg/NLPDataSet (accessed on 15 June 2025)). 1 The goal of the NLI task is to determine the logical relationship between hypothesis and premise, including three categories of relationships: entailment, neutral, and contradiction.
Table 2. Accuracy (%) of CM comparison in each dataset.
Table 2. Accuracy (%) of CM comparison in each dataset.
Dataset r = 00.10.20.30.40.50.60.70.80.9
OCNLI69.5968.8566.2062.6054.2449.3741.2834.4226.8022.97
CMNLI79.9979.9479.2379.0376.9461.9934.9437.3023.2917.27
CINLI91.1491.3190.7689.0487.1377.7079.0575.5348.7048.70
Table 6. OCNLI results under the optimal reduction ratio ( r = 0.1).
Table 6. OCNLI results under the optimal reduction ratio ( r = 0.1).
Data ProcessingAccuracyPrecisionRecallF1
Base68.9570.2268.9569.08
Sort69.7670.4969.7669.91
Sort & Reducing 10%68.2870.3168.2868.48
Table 7. Mean ± standard deviation of multiple runs on the OCNLI dataset.
Table 7. Mean ± standard deviation of multiple runs on the OCNLI dataset.
Data ProcessingAccuracyPrecisionRecallF1
Base69.15 ± 0.2970.19 ± 0.1569.15 ± 0.2969.23 ± 0.31
Sort70.32 ± 0.4871.02 ± 0.5470.32 ± 0.4970.45 ± 0.47
Sort & Reducing 10%68.69 ± 0.6370.25 ± 0.3468.69 ± 0.6368.88 ± 0.64
Table 8. CINLI results under the optimal reduction ratio ( r = 0.3).
Table 8. CINLI results under the optimal reduction ratio ( r = 0.3).
Data ProcessingAccuracyPrecisionRecallF1
Base91.785291.787391.785291.7861
Sort91.867691.867791.867691.8651
Sort & Reducing 30%90.422390.465590.422390.3838
Table 9. Mean ± standard deviation of multiple runs on the CINLI dataset.
Table 9. Mean ± standard deviation of multiple runs on the CINLI dataset.
Data ProcessingAccuracyPrecisionRecallF1
Base91.67 ± 0.1191.67 ± 0.1291.67 ± 0.1191.66 ± 0.12
Sort91.81 ± 0.0591.81 ± 0.0591.81 ± 0.0591.81 ± 0.04
Sort & Reducing 30%90.07 ± 0.4190.13 ± 0.3990.07 ± 0.4190.03 ± 0.41
Table 10. Hypothesis information statistics of CMNLI.
Table 10. Hypothesis information statistics of CMNLI.
LabelMinMaxMedianMean
01.099.016.017.1
11.0100.017.018.3
21.087.015.016.1
Table 11. The proportions of assumptions in different length intervals of CMNLI.
Table 11. The proportions of assumptions in different length intervals of CMNLI.
Label≤10 Tokens11–15 Tokens16–20 Tokens21–25 Tokens26–30 Tokens≥31 Tokens
00.2010.2790.2440.1450.0710.060
10.1290.2770.2720.1680.0840.070
20.2250.3090.2410.1280.0560.041
Table 12. Hypothesis information statistics of OCNLI.
Table 12. Hypothesis information statistics of OCNLI.
LabelMinMaxMedianMean
02.054.010.010.7
13.060.011.011.9
22.055.010.011.0
Table 13. The proportions of assumptions in different length intervals of OCNLI.
Table 13. The proportions of assumptions in different length intervals of OCNLI.
Label1–5 Token6–10 Token11–15 Token16–20 Token≥21 Token
00.0830.4790.3040.0950.039
10.0460.4250.3370.1260.066
20.0620.4750.3210.1020.040
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, F.; Zhou, W. Quality over Quantity: An Effective Large-Scale Data Reduction Strategy Based on Pointwise V-Information. Electronics 2025, 14, 3092. https://doi.org/10.3390/electronics14153092

AMA Style

Chen F, Zhou W. Quality over Quantity: An Effective Large-Scale Data Reduction Strategy Based on Pointwise V-Information. Electronics. 2025; 14(15):3092. https://doi.org/10.3390/electronics14153092

Chicago/Turabian Style

Chen, Fei, and Wenchi Zhou. 2025. "Quality over Quantity: An Effective Large-Scale Data Reduction Strategy Based on Pointwise V-Information" Electronics 14, no. 15: 3092. https://doi.org/10.3390/electronics14153092

APA Style

Chen, F., & Zhou, W. (2025). Quality over Quantity: An Effective Large-Scale Data Reduction Strategy Based on Pointwise V-Information. Electronics, 14(15), 3092. https://doi.org/10.3390/electronics14153092

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop