Next Article in Journal
Enhanced Control of Shunt Active Power Filter for Non-Active Current Compensation and Power Management in DC-Powered Systems
Previous Article in Journal
Compact Microcontroller-Based LED-Driven Photoelectric System for Accurate Photoresponse Mapping Compatible with Internet of Things
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Feature Selection Method Based on a Convolutional Neural Network for Text Classification

1
International Business School, Guangdong University of Finance and Economics, Guangzhou 510320, China
2
Department of Electronic Business, South China University of Technology, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(23), 4615; https://doi.org/10.3390/electronics14234615
Submission received: 25 October 2025 / Revised: 20 November 2025 / Accepted: 22 November 2025 / Published: 24 November 2025

Abstract

Feature selection, the technique to distinguish discriminative terms among large quantities of irrelevant and noisy terms in high-dimensional text data, is the effective solution for reducing computational cost and improving the performance of text classification. To address the issues of effective feature selection in text classification, a new ranking feature selection method is developed by utilizing the deep learning model convolutional neural network (CNN). Our method uses a pre-trained three-layer CNN to capture deep features of terms and selects discriminative terms according to the combination of deep features and term frequencies, aiming to improve performance of feature selection for text classification. Moreover, the CNN training in our method is relatively fast because the loss function of the CNN does not need to reach convergence. Nine benchmark datasets and several comparative methods are introduced to validate the effectiveness of our method. Experimental results demonstrate the following: (1) Our method achieves better performance than other comparative methods in improving text classification accuracy, with significance level of 0.05 in most cases. (2) The top 20 terms selected by our method are more semantically related to the topics of datasets than comparative methods. (3) Our method outperforms comparative methods in reducing the sparsity of text. The sparsity reduction effect generally falls in [2%, 8%].

1. Introduction

Text classification, the task of classifying text documents with unknown class labels according to predefined categories [1], is significant for further mining of the high commercial value [2] contained in text because it organizes the text documents into obvious classes. The applications of text classification cover various areas including public security [3], health care [4], proactive personality prediction [5], classification of job advertisements [6], and classification of patent documents [7]. Due to the potential application in various scenarios, text classification is still a hot spot for researchers. The research directions include (1) optimized classification models based on single deep neural networks [8,9] or hybrid deep neural networks [10,11,12,13]; (2) improved traditional classification models such as naïve Bayes [14]; (3) hybrid classification models based on traditional models and deep neural networks [15]; (4) classification models based on other models such as topic models [16]; (5) optimized text representation methods [17] to improve performance of classification models; and (6) auxiliary information, such as information on logical relationships in text data [18], similarity based on graph and semantic knowledge [19], word importance [20], and embeddings of labels [21] to improve text classification performance.
The high-dimensional text data usually contains large quantities of irrelevant and noisy terms (non-discriminative terms). The non-discriminative terms cause steep computational complexity in learning processes and have negative effects on text classification performance. Feature selection is the effective technique to distinguish discriminative terms from non-discriminative ones before the task of text classification so that the dimension of text is reduced efficiently and the classification accuracy is improved effectively. Previous feature selection methods for the text domain select terms according to only semantic information such as class labels and term (document) frequencies. Recently, deep neural networks have been successfully applied in various domains including natural language processing. According to [22], deep neural networks are able to create deep feature representations through a series of complex transformations. Deep information is a kind of useful information for data mining. We think that deep information can be integrated with semantic information to improve the performance of feature selection for text classification. Taking advantage of effective deep information, this paper proposes a new ranking feature selection method termed CNNFS (CNN-based Feature Selection). Our method is developed by using a convolutional neural network (CNN), a kind of classical deep neural network. CNNFS uses a pre-trained three-layer CNN to extract deep information of terms and performs feature selection by integrating deep information and semantic information.
Our main contribution is to illustrate an effective feature selection method, CNNFS. On the one hand, CNNFS applies a deep learning model CNN to extract deep information of terms and select discriminative terms by evaluating the integration of term frequencies and deep information. Existing studies have not considered designing effective feature selection methods by using deep information. On the other hand, CNNFS outperforms comparative feature selection methods in experiments, including improving the accuracy of text classification, selecting high proportions of semantic terms, and reducing the sparsity of text.
The other parts of this paper are organized into the following sections: Section 2 discusses previous feature selection methods in the text domain and applications of CNNs for text mining. The methodology of our method, CNNFS, is illustrated in Section 3. Section 4 covers all the details of experiments for validating the effective performance of CNNFS. Lastly, the conclusions of this paper are given in Section 5.

2. Literature Review

2.1. Feature Selection Methods in the Text Domain

Previous feature selection methods in the text domain can be categorized as ranking methods and selecting methods. Ranking methods rank terms according to ranking functions and select the top N terms to construct a discriminative subset ( N is determined manually according to experiments or practical demands). Selecting methods determine the discriminative subset of terms by other algorithms automatically, without setting values of N .

2.1.1. Ranking Methods

Various methods evaluate relationships between terms and classes. Bahassine et al. [23] introduced an improved version of Chi-square (CHI), for the purpose of balancing the selected terms from each class. Cekik and Uysal [24] designed a new method based on the concept of rough sets. Zhou et al. [25] combined term frequencies and document frequencies in both classes and datasets to design a new method. Amazal and Kissi [26] proposed a method based on MI, variances of frequencies of terms in classes, and maximum frequencies of terms. Parlak and Uysal [27] proposed a new method which performed feature selection according to the integration of conditional probabilities of terms given classes and conditional probabilities of classes given terms. Cekik and Uysal [28] presented a feature selection which selected discriminative terms by projecting the relationship among a term and any two of the classes to a two-dimensional integer space. Parlak [29] designed a new method by using conditional probabilities of terms given classes and joint probabilities of terms and classes. Jin et al. [30] proposed three new feature selection methods. The three methods were, respectively, designed based on term frequencies in classes, document frequencies in classes, and the integration of the two. Verma and Sahu [31] proposed a method for multi-label text classification. The method selected features based on relevance between features and labels. Ige and Gan [32] proposed a two-objective method which selected features by ensemble filters and then reduced the quantity of the selected features. Mohanrasu et al. [33] developed a ridge-regression-based method which evaluated relationship between features by a relation matrix. Cekik [34] introduced a method based on the knowledge of class, document, and document–class. Farek and Benajdja [35] designed a method which considered mutual information and semantic correlation. Liu et al. [36] developed several improved versions of feature selection methods based on frequency and probability.
There are methods which perform feature selection by term frequencies or document frequencies. Ashokkumar et al. [37] developed a new method which combined the information of TF-IDF (term frequency–inverse document frequency), weights of nouns, and the reciprocal of the total number of terms in documents. Garg [38] introduced a feature selection method based on a co-occurrence term network. A node of the network represented a term, while an edge represented the relationship between two adjacent terms. Okkalioglu [39] proposed a method based on the distances between terms.
Some studies pay attention to redundancy of terms and explanation. Lazhar and Amira [40] developed a method based on the assumption that semantically related terms lead to redundancy. Rehman et al. [41] proposed a method to address the problem of explanation of CNNs.

2.1.2. Selecting Methods

Selecting methods apply various kinds of methods to select discriminative subsets of terms, instead of setting the parameter N manually.
Evolutionary algorithms can be used for selecting discriminative subsets of terms. Chantar et al. [42] proposed an improved binary grey wolf optimizer and used it to design a new feature selection method. Alalyani and Larabi-Marie-Sainte [43] proposed a new method based on the firefly algorithm. Thirumoorthy and Muneeswaran [44] used a hybrid poor and rich optimization algorithm to select optimal subsets in their feature selection method. Hosseinalipour et al. [45] proposed two methods which were developed based on two versions of the farmland fertility algorithm. Adel et al. [46] proposed a new method based on the binary bat algorithm. Priya and Karthika [47] developed an embedded method to generate an optimal feature subset and then to determine the importance of features. Singh et al. [48] proposed a method for improving the global solution convergence. Wu et al. [49] designed a new method based on their proposed binary black hole algorithm. Kaya et al. [50] integrated migrating birds optimization feature selection with text classifiers to improve accuracy and efficiency. Msallam et al. [51] introduce three binary-fire-hawk-based methods for unsupervised feature selection. Nachaoui et al. [52] developed an improved PSO (particle swarm optimization) method which incorporated several components including weight, constriction factor, chaotic map, and a fitness function for penalty. Dhal and Azad [53] proposed a two-stage method which performed filter feature selection and then multi-objective wrapper feature selection. Farek and Benaidja [54] proposed an improved binary PSO method based on mutual information and a local search method.
Discriminative subsets of terms can also be selected by various other methods. El-Hajj and Hajj [55] introduced a method which transformed feature selection to an optimal problem. The method selected discriminative terms by maximizing the difference between classes and minimizing the difference between terms in the same class. Saeed and Al Aghbari [56] proposed a method based on association rules to generate discriminative subsets of terms for each class. Similarly, Farghaly and Abd El-Hafeez [57] presented a method which used the Apriori algorithm to recognize discriminative subsets of terms for each class. Sagbas [58] introduced a method which generated subsets by using a greedy search on ranked features. Liu et al. [59] proposed a two-stage method for text classification, based on the designs of multiple objects and multiple filters.

2.1.3. Summary

Both ranking methods and selecting methods perform feature selection mainly according to supervised information (relationships between terms and classes) and unsupervised information (e.g., document frequencies and term frequencies). These two kinds of information are semantic information. Supervised information is helpful to evaluate the significance of terms under classes. However, supervised information is hard to obtain because most text documents are not assigned class labels in practical scenarios. Moreover, supervised information may bring negative effects on feature selection in imbalanced datasets because the selected subsets of terms may contain more information of frequent classes. Unsupervised information is naturally contained in text documents, but it is limited because some important information (e.g., position information of terms in sentences) of text documents is filtered out.
Ranking methods rank terms by evaluation functions and select subsets of terms by using the manual parameter N . Ranking methods have two advantages. On the one hand, the evaluation of each term can be performed separately. One the other hand, the sizes of discriminative subsets of terms can be controlled manually, according to practical necessity. In the aspect of core ideas, selecting methods in our literature review can be divided into two types. One of the two types ranks terms and selects discriminative subsets using other algorithms (e.g., probability models or optimization algorithms) according to the ranked terms. The other type selects discriminative subsets directly by using optimization algorithms. Compared with ranking methods, selecting methods create much computational complexity. Moreover, most of the selecting methods are not flexible to control the sizes of discriminating subsets. In summary, ranking methods are more flexible and efficient than selecting methods.

2.2. CNNs for Text Mining

CNNs have been widely used in tasks related to images, including image compression [60], image classification [61], and detection of image forgery [62]. At the same time, CNNs have also been used in the tasks of text representation, text classification, and semantic analysis such as sentiment analysis and opinion analysis.

2.2.1. Text Representation

CNNs have been used as the tool for achieving text representation in previous studies. Wang et al. [63] developed a complex CNN with probability distribution settings for text representation. Xu et al. [64] designed a text representation solution by integrating a CNN and a variational autoencoder. Gasmi et al. [65] conducted the task of effectively retrieving medical images by using a CNN to achieve representation of queries. Liu et al. [66] and Wahyu Trisna et al. [67] utilized a CNN to discover pattens to better understand a text.

2.2.2. Text Classification

CNNs can be used as text classifiers. Wu et al. [68] developed a CNN which could capture context information for text classification. Xu et al. [69] developed a CNN classification model which applied context information of words as extra input data. Alsaleh and Larabi-Marie-Sainte [43] proposed a classifier based on GA and a CNN. GA was applied to optimize the parameters of the CNN. Butt et al. [70] proposed an improved CNN model. This model was mainly constructed by multiple convolution layers, a maxpooling layer, an averaging pooling layer, and a fully connected layer. Liang et al. [71] proposed a CNN with multiple attention settings for text classification. Liu et al. [72] developed a CNN classifier with multiple convolution layers and multiple pooling layers. Zhao et al. [73] also proposed a similar CNN classifier. Qorich and EL Quazzani [74] used a CNN structure of three convolution layers, a pooling layer, and three fully connected layers for text sentiment classification. Thekkekara et al. [75] developed a classification model for detecting depression. The model was constructed by combining a CNN, long short-term memory (LSTM), and an attention mechanism. Guo et al. [76] utilized TextCNN as a student model for their proposed distillation model.
CNNs can also be used to extract deep features of terms for text classification. Liu et al. [77] developed a complex classifier which used a CNN to extract a representation of input data. Zeng et al. [78] proposed a CNN with several kernels to process input text data. Lyu and Liu [79] introduced a text classification model based on a recurrent neural network (RNN) and a CNN. In this model, the CNN was used for capturing the importance of words. Wang et al. [80] used a CNN with an attention mechanism to discover related words in text data. Liu et al. [81] used a graph CNN to capture sentiment features of text in their proposed classification model. Zhou et al. [82] used a CNN framework to capture semantic information at multiple levels for text classification. Xiong et al. [83] and Li et al. [84] utilized a CNN to extract local features from text.

2.2.3. Semantics Analysis

CNNs have been applied in sentiment analysis. Huang et al. [85] predicted sentiment strength by using a CNN to analyze three types of embeddings of words. Krishnan et al. [86] conducted sentiment analysis by using a CNN which was improved by an optimization algorithm. Usama et al. [87] used a CNN to capture high-level features from input data. Wang et al. [88] integrated two types of CNN for their sentiment analysis model. One CNN was used to analyze context information, while the other one was used for capturing sentiment information. Ghorbanali et al. [89] proposed an ensemble network of multiple CNNs for multimodal sentiment analysis. Huang et al. [90] developed a model which analyzed embedding of sentences and sentiment words by using parallel CNN structures. Murugaiyan and Uyyala [91] proposed a serial CNN structure for recognizing emotion from speeches. Mutinda et al. [92] designed a model for sentiment analysis by integrating BERT and CNN. Alnowaiser [93] utilized a CNN to extract features for citation sentiment analysis. He and Abisado [94] proposed a model for sentiment analysis on short film comments. The model utilized a CNN to generate local features of word vectors.
CNNs have also been involved in various other tasks of semantics analysis. Chen et al. [95] applied a CNN to detect verbal aggression on Twitter. Qiu et al. [96] presented a model based on a CNN for automatically discriminating legal documents. Heo et al. [97] proposed a model for predicting atrial fibrillation. The model used a CNN to extract features. Li and Yin [98] compared the performance of different models based on a CNN and RNN in the task of analyzing semantic features. Jian et al. [99] proposed a hybrid model for measuring the readability of English text. The model was designed based on a CNN, RNN, and attention mechanism. Qiu et al. [100] proposed a CNN structure with a mechanism for filtering words for extracting spatial relation from text. Boukhers et al. [101] proposed a CNN model for ICD code prediction of clinical text. The model is constructed by parallel structures of multiple convolution layers. Muppudathi and Krishnasamy [102] proposed an improved CNN model for detecting anomalies on social media. The improvement in the model reflected on the pooling operation and the activation function. Fan [103] utilized a CNN to generate local features for detecting cyberbullying. Zeng [104] developed a CNN-based model to analyze emotional semantics from text. Faseeh et al. [105] proposed a hybrid model for identifying duplicate questions. The model learned local and long-term dependencies by using a CNN and LSTM, respectively. Wu et al. [106] developed a multimodal CNN for fake news detection.

2.2.4. Summary

CNNs have been used as text classifiers or feature extractors to discover deep features of terms in text mining. However, the deep features extracted by CNNs are unable to be explicitly interpreted. The use of deep features discovered by CNNs in feature selection can fill the gap between implicit deep information and explicit semantic information. Moreover, deep features extracted by CNNs are also the supplementary information for the limited information contained in term frequencies.

2.3. Research Motivation

This paper is motivated by the following discussions on related work.
Ranking methods are more flexible than selecting methods. Ranking methods are flexible and efficient at evaluating terms separately and controlling the dimensions of discriminative subsets conveniently. Compared with ranking methods, some of the selecting methods must rank terms by using ranking methods before selecting discriminative subsets. Furthermore, selecting methods select discriminative subsets by using complex techniques such as probability models or optimization algorithms. In summary, selecting methods create much computational complexity, compared with ranking methods, so we design our method CNNFS as a ranking method.
Previous feature selection methods use supervised information or unsupervised information for feature selection. The two kinds of semantic information are limited. Firstly, supervised information is helpful to evaluate significance of terms under classes, but it is hard to obtain because most text documents in practical scenarios are not assigned class labels. Secondly, supervised information may bring negative effects on feature selection in imbalanced datasets. The selected subset of terms may contain more information about frequent classes. Last but not least, unsupervised information is naturally contained in text documents, but it is limited because semantic information (e.g., position information of terms in sentences) of text documents is filtered out, especially in the bag-of-words model. According to the above discussion, CNNFS is designed to perform feature selection by using unsupervised information.
Deep information extracted from terms by a CNN can be utilized for improving the performance of feature selection. CNNs have been demonstrated as an effective deep model to extract deep features from images and have also been used in text mining. Deep information extracted by CNNs can be useful supplementary information for the extension of limited supervised information or unsupervised information to achieve better feature selection performance. Based on the above discussion, CNNFS integrates unsupervised information and deep information extracted by a CNN to perform feature selection.

3. Methodology

This section discusses the methodology of this paper. Section 3.1 presents the research objectives of this paper based on the discussion in Section 2.3. Section 3.2 introduces the structure of our three-layer CNN. The structure of our CNN is specially designed for selecting discriminative terms in text. Section 3.4 illustrates our methodology to use the three-layer CNN for feature selection.

3.1. Research Objectives

According to Section 2.3, we determine the following four research objectives.
Objective 1: We aim to design a ranking method, CNNFS. A ranking method is flexible to evaluate the effectiveness of terms separately and control the size of discriminative subsets of terms.
Objective 2: CNNFS performs feature selection according to the combination of unsupervised term frequency information and deep information extracted by a CNN to achieve better performance.
Objective 3: CNNFS trains the CNN by using a flexible strategy to emphasize the efficiency of feature selection.
Objective 4: CNNFS achieves good performance in improving text classification accuracy, recognizing high proportion of semantic terms, and reducing sparsity of text.
Objectives 1 to 3 are achieved in the design of our method, CNNFS (Section 3.4). Objective 4 is achieved in experiments (Section 4).

3.2. Feature Selection Preliminary

This section describes the task of feature selection mathematically. Let D = { D i | i = 1 , 2 , , | D | } and T = { T j | j = 1 , 2 , , | T | } denote a set containing | D | text documents and a set containing | T | terms, respectively. All the terms in T belong to the text documents in D . Under the bag-of-words model, text documents in D are transformed to a | D | × | T | document–term matrix (shown in Figure 1). The i th text document is represented as D i = [ t i 1 , t i 2 , , t i | T | ] T . The j th term in T is represented as T j = [ t 11 , t 21 , , t | D | j ] . t i j represents the frequency of term T j in document D i . For text classification, feature selection is responsible for the task of identifying discriminative terms from T to construct a small subset T ( T T , | T | | T | ) . Classifiers are able to achieve better performance by learning from T instead of T because T contains many fewer non-discriminative terms. Moreover, the training complexity is much lower because T is much smaller than T .

3.3. Structure of the Three-Layer CNN

CNNFS trains a three-layer CNN to discover deep features in T j . The structure of the CNN is shown in Figure 1. The CNN accepts T j as input and outputs the deep feature F ( T j ) . The three layers, namely, the smoothing layer, extending layer, and compressing layer, are specially designed to perform different processes for T j . The smoothing layer is designed for smoothing T j and reducing its sparsity of term frequencies. The extending layer is designed for extending information of T j . The compressing layer is designed for further processing and compressing the output of extending layer so that the dimensions of T j and F ( T j ) are the same. Table 1 summarizes the CNN architecture and training parameters.

3.3.1. Basic Operations

There are three kinds of operations involved in our CNN structure: the p × p padding, the k × k kernel, and the s × s strides. These operations are used to adjust values of term vectors and control the dimensions of the output deep features.

3.3.2. Basic Layers

(1)
Smoothing Layer
The smoothing layer is the first layer of our CNN, and it contains a convolution layer and a maxpooling layer. The input of this layer is the term vector T j = [ t 11 , t 21 , , t | D | j ] , and the output is the feature vector of T j . The convolution layer and the maxpooling layer are used for discovering deep information of T j . Furthermore, the convolution layer can also ease the problem of sparsity of T j . Figure 2 illustrates structure of the smoothing layer. The convolution layer first processes the input D × 1   T j and creates a D × 1 feature. Next, the maxpooling layer processes the feature and creates the final feature S L ( T j ) . The * operator is the convolution operator.
(2)
Extending Layer
The extending layer (shown in Figure 3) is constructed by four independent smoothing layers. Each smoothing layer processes the input D × 1   S L ( T j ) and creates an output D × 1   S L ( S L T j ) . Four of the four D × 1   S L ( S L T j ) s are concatenated by rows to create two D × 2   S L ( S L T j ) s. The two D × 2   S L ( S L T j ) s are further concatenated by columns to create the final output, 2 D × 2   S L ( S L T j ) . The extending layer aims to discover further deep information about S L ( T j ) by the smoothing layer. The concatenation operation is used to create more variation and offer more relationships among elements in T j for the up-coming compressing layer to discover more deep information. The * operator is the convolution operator.
(3)
Compressing Layer
The final compressing layer (shown in Figure 4) processes the 2 D × 2   S L ( S L T j ) . The convolution layer is used to discover more deep information from 2 D × 2   S L ( S L T j ) . The maxpooling layer is used to compress the 2 D × 2 output of the convolution layer to a final D × 1 deep feature. Moreover, the maxpooling layer also reduces the sparsity of features to some extent. The D × 1 deep feature is further processed by a fully connected layer to create the final deep feature, F ( T j ) . We can see that F ( T j ) is of the same dimensions as T j . The * operator is the convolution operator.

3.4. The Proposed Feature Selection Method, CNNFS

The flow chart of feature selection using CNNFS is shown in Figure 5. CNNFS performs feature selection in two stages: (1) First stage, a three-layer CNN (shown in Figure 3) is trained. (2) Second stage, discriminative terms are selected according to the trained CNN in the first stage.
(1)
First Stage
Each T j = [ t 11 , t 21 , , t | D | j ] in the document–term matrix is input one by one to train the CNN. The output of the CNN is a deep feature of T j : F T j = [ f 11 , f 21 , , f | D | j ] . F T j T j 2 evaluates the difference between T j and F T j . A large value of F T j T j 2 indicates that T j contains a lot of deep information.
(2)
Second Stage
In this stage, each T j is input to the trained CNN to achieve its deep feature F T j . The score of T j   C N N F S ( T j ) is calculated according to Formula (1). As mentioned above, a large value of F T j T j 2 indicates that T j contains a lot of deep information, so terms with large values of C N N F S ( T j ) are considered discriminative. The deep information is reflected in F T j .
C N N F S ( T j ) = F T j T j 2
(3)
Algorithm
The process of performing CNNFS for feature selection is illustrated by the algorithm shown in Algorithm 1. The first stage starts at line 2 and stops at line 6. Function i n i t i a l i z e C N N ( ) initializes the CNN in Figure 3. Line 3 to line 6 trains the CNN using T j s. The CNN training stops at j = | D | . Our training strategy has three advantages in simplifying the training process and reducing training time. Firstly, the stop point ( j = | D | ) of CNN training is definite. Secondly, each term vector only needs to be input once. Finally, the convergence of the loss function is unnecessary. Line 7 to line 12 shows the second stage. F T j is calculated for each T j by the trained CNN. The scores of T j are stored in F S , the vector of length | T | . Elements in F S are then sorted in decreasing order. The output subset T contains the top N discriminative terms in F S . N is a manual parameter which is set according to experiments or practical needs.
Algorithm 1. Algorithm of feature selection using CNNFS
Input:  d t m , the | D | × | T | document-term matrix
N , the number of selected discriminating terms
Output:  T , the subset of discriminating terms
1
2
3
4
5
6
7
8
9
10
11
12
13
14
B e g i n
 # first stage
C N N = i n i t i a l i z e C N N ( ) ;
For   j = 1 | D |
T r a i n C N N ( T j , C N N ) ;
End For
 # second stage
F S = v e c t o r ( l e n g t h = | T | ) ;
For   j = 1 | D |
F T j =   C N N ( T j ) ;
F S T j = F T j T j 2 ;
End For
 # output results
T = s o r t F S ,   d e c r e a s i n g = T r u e [ 1 : N ] ;
15Return  T ;
16 E n d
(4)
Complexity
According to [107], the complexity of a certain number of traditional supervised and unsupervised filter methods is O ( D × | T | )   +   O [ s ( | T | ) ] . | D | represents the number of text documents. | T | represents the total number of unique words in the | D | documents. O ( D × | T | ) represents the complexity of scoring the | T | unique words. O [ s ( | T | ) ] represents the complexity of sorting the | T | unique words according to their scores. According to Algorithm 1, the complexity of CNNFS is also O ( D × | T | )   +   O [ s ( | T | ) ] . Moreover, CNNFS has the advantages of utilizing more deep information than traditional filter methods.

4. Experiments

4.1. Datasets

Nine benchmark datasets are applied for comparison experiments. The details of these nine datasets are illustrated in Table 2. The first two datasets (CF and CR), respectively, contain user favorite description text and user review text of 12 different Toyota cars. They are generated from the original OpinRank Review dataset, which is applied in [108]. CNAE, utilized in [109,110], consists of business description texts about Brazilian companies, which are categorized into nine groups. IMDB, consisting of positive and negative movie reviews, is one of the subsets in the Sentiment Labeled Sentences Dataset in [111]. KDC and TTC are datasets which contains Turkish text of articles and news. KDC is utilized for experiments in [112,113,114]. TTC is introduced in [115]. The aforementioned datasets are obtained from the famous UCI repository. The WEBKB dataset [116,117] is a collection of web pages of computer science departments of four universities in January 1997. The R8 and R52 datasets are the subsets of the Reuters-21578 dataset [116,118,119,120,121], a collection of economic news published by Reuters in 1978.
There are two reasons for choosing these datasets in our comparative experiments: (1) The diversity of different topics (e.g., cars, business, movies, computers, and news) in the datasets guarantees the robustness and generalization of the results in comparative experiments. (2) A certain number of researchers have used these datasets in experiments of several text mining tasks (e.g., text classification, opinion analysis, and information retrieval), so the reliability of these datasets can be guaranteed.

4.2. Evaluation Metrics

Support vector machine (SVM), the effective classification algorithm, and multinomial naïve Bayes (MNB), the classical text classification algorithm, are utilized to train classifiers in experiments. Classification accuracy [22] is used as an evaluation metric classifier. Tenfold cross-validation is applied in order to guarantee more accurate classification performance.
Six comparative methods (shown in Table 3) are introduced in our experiments. The NDM method was recently proposed by [122]. NDM is described in the view of binary class in the original paper, but it can be extended for the case of multi-class in the view of weighted averaging. The other five methods are usually used as the comparative methods in previous studies. All six feature selection methods select terms with high scores as discriminative terms.

4.3. Results and Discussions

This section presents experimental results and discusses these results in three different aspects: classification accuracy analysis, semantics analysis, and sparsity analysis. Classification accuracy analysis compares our CNNFS with six other methods on improving classification accuracy of SVM and MNB classifiers. Semantic analysis compares the terms selected by our method and comparative methods on the aspect of semantic meanings. Sparsity analysis compares sparsity of document–term matrices under T selected by our method and other methods.

4.3.1. Classification Accuracy Analysis

Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 show the classification accuracy of our method CNNFS and six other comparative methods. The specific classification accuracy is presented in Table A1 in Appendix A. Let N denote the number of terms. All the methods can be grouped into three levels.
In Figure 6, Level 1 contains CNNFS. Level 2 contains ECE, MI, WOR, and NDM. ECE outperforms MI. ECE outperforms WOR when N 150 for MNB. However, WOR outperforms ECE when N 250 for SVM. WOR outperforms MI when N 50 for SVM and N 250 for MNB. NDM performs worst at Level 2, while GI surpasses IG at Level 3.
In Figure 7, Level 1 contains CNNFS. Level 2 contains ECE, MI, GI, WOR, and NDM. ECE achieves the highest accuracy in Level 2 when N 200 . MI outperforms WOR when N 50 for SVM and N 100 for MNB. MI outperforms NDM when N [ 150,200,250,300,350,400,450,500 ] for SVM and N 150 for MNB. NDM outperforms WOR when N [ 100,150,200,250,300,350,500 ] for SVM and N 200 for MNB. GI achieves the lowest accuracy in Level 2. Level 3 contains IG.
In Figure 8, Level 1 contains CNNFS and ECE. CNNFS outperforms ECE. Level 2 contains MI, GI, WOR, and NDM. MI achieves similar accuracy to NDM. WOR achieves the highest accuracy in Level 2 when N 100 . GI achieves the lowest accuracy in Level 2 when N 160 . Level 3 contains IG.
In Figure 9, Level 1 contains CNNFS and WOR. WOR outperforms CNNFS when N 150 . Level 2 contains MI, GI, and NDM. MI outperforms GI and NDM. GI achieves similar accuracy to NDM. Level 3 contains ECE and IG. ECE outperforms IG when N 150 . ECE achieves similar accuracy to GI and NDM when N 240 .
In Figure 10, Level 1 contains CNNFS and ECE. CNNFS outperforms ECE. Level 2 contains MI, WOR, and NDM. WOR outperforms MI and NDM when N 200 . MI, WOR, and NDM achieve similar accuracy when N   200 . Level 3 contains IG and GI. GI outperforms IG.
In Figure 11, Level 1 contains CNNFS, and Level 2 contains ECE, MI, and NDM. ECE outperforms NDM when N 200 . MI outperforms ECE and NDM. ECE outperforms NDM when N   200 . Level 3 contains IG, GI, and WOR. IG achieves the lowest accuracy in Level 3. IG outperforms WOR when N   50 for SVM and when N   200 for MNB. IG achieves similar accuracy to WOR when N   200 .
In Figure 12, Level 1 contains CNNFS and ECE. CNNFS outperforms ECE when N   450 for SVM and N   350 for MNB. Level 2 contains MI, GI, WOR, and NDM. NDM achieves the highest accuracy in Level 2. WOR achieves the second-highest accuracy in Level 2 when N   150 . GI outperforms MI when N   50 for SVM and N   150 for MNB. Level 3 contains IG.
In Figure 13, Level 1 contains CNNFS and ECE. CNNFS outperforms ECE. Level 2 contains MI, WOR, and NDM. Generally, NDM outperforms WOR, while WOR outperforms MI. Level 3 contains IG and GI. GI outperforms IG.
In Figure 14, Level 1 contains CNNFS. Level 2 contains ECE, WOR, and NDM. ECE achieves the highest accuracy in Level 2. NDM outperforms WOR when N   150 for SVM and N   300 for MNB. Level 3 contains IG, GI, and MI. Generally, MI outperforms GI, while GI outperforms IG.
Furthermore, t-tests are conducted to statistically validate the performance of CNNFS. The p-values of the t-tests are given in Table A2 in Appendix A. In most cases (except the four cases of bold p-values in Table A2 (3) and (4)), the p-values are smaller than 0.05, indicating a significant difference between the performance of CNNFS and the performance of comparative methods.
In summary, CNNFS outperforms six other comparative methods in most of the classification cases. IG generally achieves the lowest accuracy among all the methods. The performance ECE, GI, MI, WOR, and NDM is not stable in experiments.

4.3.2. Semantics Analysis

This section presents discussions on the semantics of terms selected by CNNFS and comparative methods. For better illustration, we select semantic terms which are related to the topics of the corresponding datasets out of the top 20 terms from CF, CR, IMDB, WEBKB, R8, and R52. Terms in other datasets are unable to be recognized because they are replaced by variable names (e.g., “feature X”). The semantic terms in the top 20 terms of the datasets are summarized in Figure 15. The bold terms are the manually selected semantic terms of datasets (all the terms are stemmed, and all the uppercase letters are transformed to lowercase). The percentages in circles indicate the percentage of semantic terms in the top 20 terms which are selected by feature selection methods.
The CF dataset relates to topic of Toyota cars. The semantic terms are considered as car features or sentiment words. Among the top 20 terms selected by CNNFS, 95% of them are semantic terms. Following CNNFS, ECE also achieves relatively good performance because 45% of its selected top 20 terms are semantic terms. Other comparative methods select low percentages of semantic terms for their lists of the top 20 terms. The case of the CR dataset is similar to that of the CF dataset.
The IMDB dataset contains positive or negative opinions on movies. The semantic terms are considered as movie features and sentiment words. As we can see, 60% of the selected top 20 terms of CNNFS are related to movie features and sentiment words. In addition, WOR performs better than CNNFS. Low percentages of semantic words are provided by other comparative methods in their selected top 20 terms.
The WEBKB dataset contains text from web pages from departments of computer science from four universities. The semantic terms are considered as nouns due to complexity of the contents of the dataset. CNNFS offers 80% of the semantic terms in its selected top 20 terms. NDM selects 75% of its top 20 terms as semantic terms. The percentages offered by comparative methods are relatively low.
The R8 and R52 datasets contain economic news. The semantic terms are considered as terms related to economy. Among the selected top 20 terms, CNNFS provides 50% and 55% semantic terms, respectively, for the two datasets. ECE and WOR also perform well in the R8 dataset (50% and 55%, respectively). WOR offers a relatively high percentage (40%) of economy-related terms in the R52 dataset. Relatively, other comparative methods do not perform well in these two datasets.
In summary, the selected top 20 terms selected by CNNFS are more semantically related to the topics of datasets than comparative methods.

4.3.3. Sparsity Analysis

In this section, we evaluate the sparsity of document–term matrices, which are constructed using the terms selected by CNNFS and comparative methods. For better illustration, the number of terms is set as 500, that is, T = 500 . The results are shown in Figure 16. It is obvious that CNNFS achieves the best performance in reducing the sparsity of document–term matrices in eight of the nine datasets. Moreover, the effects of the sparsity reduction are significant. The sparsity reduction effect generally falls in the range of [2%, 8%].

5. Conclusions

This paper proposes a CNN-based feature selection method, aiming to improve text classification accuracy. Our method constructs and trains a three-layer CNN to capture deep features of terms and then selects discriminative terms by the trained CNN. Terms with a lot of deep information are considered discriminative. The results in experiments demonstrate that CNNFS achieves more stable and better performance than comparative methods.
Our method, CNNFS, has three advantages: (1) CNNFS applies a deep learning model CNN to extract deep features of terms and selects discriminative terms by means of integrating term frequencies and deep features. Deep features are another view of information used as the supplementary information for the limited information contained in term frequencies. (2) Experimental results indicate that CNNFS outperforms some existing methods (as comparative methods) in the tasks of text classification, semantic terms selection, and sparsity reduction.
The research in this paper has limitations. (1) In CCNFS, the extracted deep information by the CNN is local information in text [83,84]. Local information is not that capable of reflecting global information in text. Future studies can focus on designing hybrid-deep-neural-network-based feature selection methods to leverage both local and global information, for example, a feature selection method which utilizes a CNN to extract local information and Bi-LSTM to extract global dependency information. (2) CNNFS ignores the relationship between words. The word relationship helps to evaluate the relative importance of a word among other words in the same document. Future studies can make efforts to integrate relationship information into feature selection by tools such as dependency parsing and topic modeling. (3) CNNFS is conducted based on bag-of-words frequency information which is a kind of statistical information. Pre-trained large language models (LLMs) are helpful to achieved semantic information about words. The semantic information is also useful for selecting discriminative words. Future studies can pay attention to integrating frequency information and semantic information, for more effective feature selection.

Author Contributions

Conceptualization, J.X. and M.H.; Methodology, J.X. and M.H.; Investigation, J.X. and M.H.; Writing—original draft, J.X. and M.H.; Writing—review & editing, J.X. and M.H.; Funding acquisition, J.X. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Guangdong Philosophy and Social Sciences Fund, Grant No. GD25CTS08; Guangzhou Science and Technology Plan Project, Grant No. 2025A04J3937.

Data Availability Statement

The data used to support the findings of this study is available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Appendix A

Table A1. Classification accuracy in datasets (unit: %).
Table A1. Classification accuracy in datasets (unit: %).
(1) CF Dataset
SVMCNNIGECEGIMI_avgWORNDM
5030.919.525.521.424.329.123.1
10033.917.129.319.926.431.325.3
15036.616.230.619.527.531.827.2
2003715.332.522.830.634.328
2503714.232.122.431.634.127.8
30035.613.733.722.330.932.826.4
35035.713.533.322.332.133.626.7
40036.313.533.822.332.133.826.5
45036.413.633.422.332.233.527.1
50036.414.234.622.632.433.927.9
MNBCNNIGECEGIMI_avgWORNDM
5031.0817.82621.425.228.524.2
10036.415.230.921.428.331.525
15041.313.232.821.429.531.927.3
20041.312.135.121.430.634.129.4
25041.111.63721.433.333.930.4
30041.810.938.821.634.533.332.4
35042.511.939.421.434.733.432.3
40042.512.739.120.134.733.531.8
45042.713.139.819.234.834.933.1
50043.11340.219.135.835.432.9
(2) CR Dataset
SVMCNNIGECEGIMI_avgWORNDM
5048.619.222.419.826.224.822.5
10051.618.424.422.526.125.128.8
15054.315.323.424.234.127.829
20051.112.632.422.63328.331.8
25053.711.738.423.232.628.132
30052.29.64121.534.63032.2
35051.49.143.423.832.929.934
40051.98.143.724.233.530.931.5
45052.18.443.223.933.430.630.4
50050.77.444.724.233.629.731.8
MNBCNNIGECEGIMI_avgWORNDM
505116.722.221.626.227.221.1
10058.613.825.223.727.228.128.7
15061.111.525.123.736.529.530
20062.910.23423.736.229.435.2
25065.18.641.923.537.329.135.4
30063.77.348.924.539.732.835
35063.9752.123.840.832.636.5
40064.47.451.923.940.233.636.5
45064.67.953.824.140.233.235.9
50064.77.655.824.340.833.236.8
(3) CNAE Dataset
SVMCNNIGECEGIMI_avgWORNDM
2072.89.436.714.917.430.317.4
4081.79.975.816.919.23219.2
6083.79.681.618.921.831.921.8
8086.710.184.124.828.632.228.6
10089.311.386.926.929.83230.5
12089.611.887.629.433.231.633.4
14091.616.988.731.438.534.139
16091.819.289.440.641.635.541.8
18092.826.289.844.44337.742.6
20092.932.791.844.549.344.449.6
MNBCNNIGECEGIMI_avgWORNDM
207210.739.315.217.729.917.7
4080.610.673.417.720.531.320.5
6084.710.480.62022.530.322.5
8087.610.684.223.928.631.929.4
10090.410.687.525.631.231.631.9
1209111.289.228.733.431.734.4
14092.113.889.830.937.833.538
16092.117.590.141.14134.441.6
18093.225.190.144.242.838.143.2
20093.334.391.745.749.745.650.2
(4) IMDB Dataset
SVMCNNIGECEGIMI_avgWORNDM
3060.145.945.953.156.960.653.1
6068.842.342.355.26064.855.2
9068.3393957.358.867.657.3
12070.836.636.658.163.170.958.1
15071.5343461.264.372.361.2
18071.831.345.761.665.27461.6
21071.430.156.862.466.874.262.4
24071.229.762.66368.473.463
27072.429.16664.568.675.564.5
30073.12866.86568.575.965
MNBCNNIGECEGIMI_avgWORNDM
3062.445.745.751.652.658.851.6
6069.142425656.765.456
9069.738.538.558.460.169.158.4
12072363659.363.77259.3
1507333.533.562.564.774.162.5
1807329.543.663.165.575.463.1
21074.427.65764.467.776.464.4
24074.826.562.664.769.277.464.7
27075.426.466.366.469.978.666.4
30075.626.567.966.970.98066.9
(5) KDC Dataset
SVMCNNIGECEGIMI_avgWORNDM
5062.912.12414.616.423.616.1
10072.611.244.315.618.625.618.8
15077.21164.916.120.726.221.2
20080.610.769.116.924.926.823.6
25082.110.870.917.328.928.928.8
30083.110.574.717.830.93030.4
3508410.676.217.832.431.132
4008510.677.81832.632.433.1
45085.710.378.418.135.233.834.2
500861079.518.135.934.334.9
MNBCNNIGECEGIMI_avgWORNDM
5061.911.722.714.216.124.216
10071.411.144.314.518.625.818.4
15077.210.863.615.22126.220.5
20081.410.569.315.524.426.223.5
2508410.671162827.527.8
30085.410.275.11629.828.828.9
35086.69.976.816.131.4230.129.7
40087.49.578.51632.231.231
45088.19.179.815.934.132.632
50088.98.781.116.234.332.232.2
(6) TTC Dataset
SVMCNNIGECEGIMI_avgWORNDM
5056.115.927.217.542.115.725.3
10073.815.728.518.351.315.731.1
15077.116.131.619.955.915.638.3
20080.115.547.720.559.416.141.7
25080.615.154.921.562.616.444.3
30081.515.458.422.464.817.848.1
35082.215.660.322.769.818.352.8
40082.415.86523.770.720.354
45082.516.470.725.370.122.255.2
50083.417.873.725.971.222.556
MNBCNNIGECEGIMI_avgWORNDM
5069.715.928.815.447.615.524.3
10079.215.630.914.456.315.530
15082.715.135.216.160.415.544.2
20084.314.851.717.265.11649.8
25084.714.858.318.168.315.7553.1
30085.114.663.118.270.317.256.1
35085.714.466.918.574.417.359.5
4008614.672.321.575.519.360
4508615.576.823.476.12161.6
50086.617.480.7523.677.721.463.3
(7) WEKBE Dataset
SVMCNNIGECEGIMI_avgWORNDM
5059.43953.342.840.743.346.1
10064.538.55944.241.944.147
15066.438.26346.342.949.851.3
20069.937.866.547.64549.852.6
25072.538.267.648.344.951.654
30074.537.569.249.545.453.755.5
35076.937.572.550.646.954.155.9
40078.337.676.651.347.455.557.1
4508037.379.151.747.456.157.7
50080.436.580.552.448.856.857.9
MNBCNNIGECEGIMI_avgWORNDM
5071.437.560.443.241.84251.3
10073.836.464.34544.542.552.5
15073.735.867.447.444.950.359.4
20074.834.67049.347.950.561.4
25075.334.170.150.547.951.863.1
30076.133.370.852.148.954.264.6
35077.332.575.653.251.455.665.2
40077.4317954.151.856.166
4507830.28154.751.45766.8
50078.429.881.45652.557.867.1
(8) R8 Dataset
SVMCNNIGECEGIMI_avgWORNDM
5084.75177.351.352.861.873
10089.45180.85259.865.974.6
15092.150.682.952.761.766.875.5
2009350.784.853.462.36876.4
25093.350.286.353.562.368.578.7
30093.55087.354.462.269.480.2
35093.749.889.255.162.169.781.1
40093.75089.955.462.569.881.4
45093.849.990.455.964.569.982.2
50093.849.190.456.165.970.383.9
MNBCNNIGECEGIMI_avgWORNDM
5081.550.774.651.153.564.566.9
10089.150.479.552.261.569.569.2
15092.650.181.253.164.371.272.2
20093.349.983.953.965.172.474
25093.849.585.754.364.873.178.5
3009449.386.955.164.374.280.2
35094.249.288.956.163.974.681.4
40094.348.989.756.464.674.681.7
45094.348.990.256.866.174.783.6
50094.648.990.55767.775.285.7
(9) R52 Dataset
SVMCNNIGECEGIMI_avgWORNDM
5073.642.963.543.344.25351
10079.642.669.843.647.557.458
15082.242.472.343.954.161.364.2
2008442.374.744.355.166.269.6
2508541.976.744.555.167.472.7
30086.442.177.744.65567.773.7
3508741.978.844.855.168.775.7
4008741.879.745.155.169.377.1
45087.541.380.845.35569.377.3
50087.941.581.145.454.969.778.1
MNBCNNIGECEGIMI_avgWORNDM
5071.942.862.243.244.759.653.3
10081.242.669.643.849.465.361.3
15084.542.672.94454.869.166.3
20086.142.475.244.356.972.771.9
25087.442.278.144.556.973.773.9
30088.6427944.656.974.475.9
35088.941.880.844.956.975.278.2
40089.341.781.945.256.675.679.2
45089.741.582.945.456.575.780
50089.841.283.345.443.875.980.6
Table A2. The p-values of t-tests in datasets.
Table A2. The p-values of t-tests in datasets.
(1) CF Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM2.74 × 10−83.46 × 10−52.10 × 10−94.08 × 10−61.68 × 10−62.24 × 10−12
MNB9.64 × 10−83.25 × 10−51.57 × 10−71.06 × 10−71.68 × 10−62.23 × 10−8
(2) CR Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM1.00 × 10−90.001.89 × 10−131.80 × 10−91.05 × 10−101.57 × 10−9
MNB3.95 × 10−90.001.06 × 10−101.31 × 10−101.32 × 10−106.50 × 10−14
(3) CNAE Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM4.16 × 10−110.113.66 × 10−103.14 × 10−108.86 × 10−113.25 × 10−10
MNB2.07 × 10−100.075.91 × 10−101.86 × 10−102.61 × 10−102.16 × 10−10
(4) IMDB Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM9.80 × 10−70.001.70 × 10−72.78 × 10−50.191.70 × 10−7
MNB1.39 × 10−60.002.25 × 10−92.33 × 10−60.392.25 × 10−9
(5) KDC Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM4.80 × 10−100.001.20 × 10−107.67 × 10−134.64 × 10−115.55 × 10−13
MNB2.16 × 10−90.009.36 × 10−101.35 × 10−128.29 × 10−102.69 × 10−12
(6) TTC Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM1.78 × 10−98.74 × 10−54.42 × 10−106.71 × 10−77.34 × 10−101.29 × 10−8
MNB2.06 × 10−110.001.84 × 10−128.24 × 10−63.87 × 10−121.09 × 10−6
(7) WEKBE Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM1.97 × 10−70.001.20 × 10−81.51 × 10−81.61 × 10−91.21 × 10−8
MNB5.56 × 10−100.043.67 × 10−115.44 × 10−136.97 × 10−99.28 × 10−7
(8) R8 Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM2.29 × 10−111.12 × 10−52.42 × 10−134.52 × 10−144.13 × 10−151.34 × 10−8
MNB3.31 × 10−101.07 × 10−57.65 × 10−122.14 × 10−141.62 × 10−128.16 × 10−7
(9) R52 Dataset
CNN vs. IGCNN vs. ECECNN vs. GICNN vs. MI_avgCNN vs. WORCNN vs. NDM
SVM6.84 × 10−105.57 × 10−91.25 × 10−107.01 × 10−133.52 × 10−115.99 × 10−6
MNB2.63 × 10−99.65 × 10−87.61 × 10−108.86 × 10−98.08 × 10−121.76 × 10−6

References

  1. Joachims, T. Transductive Inference for Text Classification Using Support Vector Machines. In Proceedings of the 16th International Conference on Machine Learning, Bled, Slovenia, 27–30 June 1999; pp. 200–209. [Google Scholar]
  2. Tan, A.H. Text Mining: The State of the Art and the Challenges. In Proceedings of the Pakdd Workshop on Knowledge Disocovery from Advanced Databases, Beijing, China, 26–28 April 1999; pp. 65–70. [Google Scholar]
  3. Mujtaba, G.; Shuib, L.; Raj, R.G. Detection of Suspicious Terrorist Emails Using Text Classification: A Review. Malays. J. Comput. Sci. 2019, 31, 271–299. [Google Scholar] [CrossRef]
  4. Srivastava, S.K.; Singh, S.K.; Suri, J.S. Effect of Incremental Feature Enrichment on Healthcare Text Classification System: A Machine Learning Paradigm. Comput. Methods Programs Biomed. 2019, 172, 35–51. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, P.; Yang, Y.; Si, Y.D.; Zhu, G.C.; Zhan, X.P.; Wang, J.; Pan, R.S. Classification of Proactive Personality: Text Mining Based on Weibo Text and Short-Answer Questions Text. IEEE Access 2020, 8, 97370–97382. [Google Scholar] [CrossRef]
  6. Ostrogonac, S.J.; Rastovic, B.S.; Popovic, B. Automatic Job Ads Classification, Based on Unstructured Text Analysis. ACTA Polytech. Hung. 2021, 18, 191–204. [Google Scholar] [CrossRef]
  7. Li, R.; Yu, W.K.; Liu, Y.Y. Patent Text Classification Based on Deep Learning and Vocabulary Network. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 54–61. [Google Scholar] [CrossRef]
  8. Kim, J.; Jang, S.; Park, E.; Choi, S. Text Classification Using Capsules. Neurocomputing 2020, 376, 214–221. [Google Scholar] [CrossRef]
  9. Dai, Y.; Shou, L.J.; Gong, M.; Xia, X.L.; Kang, Z.; Xu, Z.L.; Jiang, D.X. Graph Fusion Network for Text Classification. Knowl.-Based Syst. 2022, 236, 107659. [Google Scholar] [CrossRef]
  10. Zhan, Z.Q.; Hou, Z.F.; Yang, Q.C.; Zhao, J.Y.; Zhang, Y.; Hu, C.J. Knowledge Attention Sandwich Neural Network for Text Classification. Neurocomputing 2020, 406, 1–11. [Google Scholar] [CrossRef]
  11. Tezgider, M.; Yildiz, B.; Aydin, G. Text Classification Using Improved Bidirectional Transformer. Concurr. Comput.-Pract. Exp. 2022, 34, e6486. [Google Scholar] [CrossRef]
  12. Xu, Y.H.; Yu, Z.W.; Cao, W.M.; Chen, C.L.P. Adaptive Dense Ensemble Model for Text Classification. IEEE Trans. Cybern. 2022, 52, 7513–7526. [Google Scholar] [CrossRef]
  13. Wang, Y.Z.; Wang, C.X.; Zhan, J.Y.; Ma, W.J.; Jiang, Y.C. Text FCG: Fusing Contextual Information via Graph Learning for text classification. Expert Syst. Appl. 2023, 219, 119658. [Google Scholar] [CrossRef]
  14. Gan, S.F.; Shao, S.Q.; Chen, L.; Yu, L.J.; Jiang, L.X. Adapting Hidden Naive Bayes for Text Classification. Mathematics 2021, 9, 2378. [Google Scholar] [CrossRef]
  15. Shi, Z.; Fan, C.J. Short Text Sentiment Classification Using Bayesian and Deep Neural Networks. Electronics 2023, 12, 1589. [Google Scholar] [CrossRef]
  16. Rijcken, E.; Kaymak, U.; Scheepers, F.; Mosteiro, P.; Zervanou, K.; Spruit, M. Topic Modeling for Interpretable Text Classification from EHRs. Front. Big Data 2022, 5, 846930. [Google Scholar] [CrossRef]
  17. Sinoara, R.A.; Camacho-Collados, J.; Rossi, R.G.; Navigli, R.; Rezende, S.O. Knowledge-enhanced Document Embeddings for Text Classification. Knowl.-Based Syst. 2019, 163, 955–971. [Google Scholar] [CrossRef]
  18. Wang, H.Y.; Zeng, D.H. Fusing Logical Relationship Information of Text in Neural Network for Text Classification. Math. Probl. Eng. 2020, 2020, 5426795. [Google Scholar] [CrossRef]
  19. Shanavas, N.; Wang, H.; Lin, Z.W.; Hawe, G. Knowledge-driven Graph Similarity for Text Classification. Int. J. Mach. Learn. Cybern. 2021, 12, 1067–1081. [Google Scholar] [CrossRef]
  20. Sulaimani, S.A.; Starkey, A. Short Text Classification Using Contextual Analysis. IEEE Access 2021, 9, 149619–149629. [Google Scholar] [CrossRef]
  21. Tan, C.G.; Ren, Y.; Wang, C. An Adaptive Convolution with Label Embedding for Text Classification. Appl. Intell. 2023, 53, 804–812. [Google Scholar] [CrossRef]
  22. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining: Practical Machine Learning Tools and Techniques; Green, T., Ed.; Morgan Kaufmann: Burlington, NJ, USA, 2017. [Google Scholar]
  23. Bahassine, S.; Madani, A.; Al-Sarem, M.; Kissi, M. Feature Selection Using an Improved Chi-square for Arabic Text Classification. J. King Saud Univ.-Comput. Inf. Sci. 2020, 32, 225–231. [Google Scholar] [CrossRef]
  24. Cekik, R.; Uysal, A.K. A Novel Filter Feature Selection Method Using Rough Set for Short Text Data. Expert Syst. Appl. 2020, 160, 113691. [Google Scholar] [CrossRef]
  25. Zhou, H.F.; Ma, Y.M.; Li, X. Feature Selection Based on Term Frequency Deviation Rate for Text Classification. Appl. Intell. 2020, 51, 3255–3274. [Google Scholar] [CrossRef]
  26. Amazal, H.; Kissi, M. A New Big Data Feature Selection Approach for Text Classification. Sci. Program. 2021, 2021, 6645345. [Google Scholar] [CrossRef]
  27. Parlak, B.; Uysal, A.K. A Novel Filter Feature Selection Method for Text Classification: Extensive Feature Selector. J. Inf. Sci. 2021, 49, 59–78. [Google Scholar] [CrossRef]
  28. Cekik, R.; Uysal, A.K. A New Metric for Feature Selection on Short Text Datasets. Concurr. Comput.-Pract. Exp. 2022, 34, e6909. [Google Scholar] [CrossRef]
  29. Parlak, B. Class-index Corpus-index Measure: A Novel Feature Selection Method for Imbalanced Text Data. Concurr. Comput.-Pract. Exp. 2022, 34, e7140. [Google Scholar] [CrossRef]
  30. Jin, L.B.; Zhang, L.; Zhao, L. Feature Selection Based on Absolute Deviation Factor for Text Classification. Inf. Process. Manag. 2023, 60, 103251. [Google Scholar] [CrossRef]
  31. Verma, G.; Sahu, T.P. Deep Label Relevance and Label Ambiguity Based Multi-label Feature Selection for Text Classification. Eng. Appl. Artif. Intell. 2025, 148, 110403. [Google Scholar] [CrossRef]
  32. Ige, O.P.; Gan, K.H. Ensemble Filter-Wrapper Text Feature Selection Methods for Text Classification. CMES-Comput. Model. Eng. Sci. 2025, 141, 1847–1865. [Google Scholar] [CrossRef]
  33. Mohanrasu, S.S.; Janani, K.; Rakkiyappan, R. A COPRAS-based Approach to Multi-Label Feature Selection for Text Classification. Math. Comput. Simul. 2024, 222, 3–23. [Google Scholar] [CrossRef]
  34. Cekik, R. A New Filter Feature Selection Method for Text Classification. IEEE Access 2024, 12, 139316–139335. [Google Scholar] [CrossRef]
  35. Farek, L.; Benaidja, A. A Non-redundant Feature Selection Method for Text Categorization Based on Term Co-occurrence Frequency and Mutual Information. Multimed. Tools Appl. 2024, 83, 20193–20214. [Google Scholar] [CrossRef]
  36. Liu, X.; Wang, S.; Lu, S.Y.; Yin, Z.T.; Li, X.L.; Yin, L.R.; Tian, J.W.; Zheng, W.F. Adapting Feature Selection Algorithms for the Classification of Chinese Texts. Systems 2023, 11, 483. [Google Scholar] [CrossRef]
  37. Ashokkumar, P.; Shankar, G.S.; Srivastava, G.; Maddikunta, P.K.R.; Gadekallu, T.R. A Two-stage Text Feature Selection Algorithm for Improving Text Classification. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2021, 20, 49. [Google Scholar]
  38. Garg, M. UBIS: Unigram Bigram Importance Score for Feature Selection from Short Text. Expert Syst. Appl. 2022, 195, 116563. [Google Scholar] [CrossRef]
  39. Okkalioglu, M. A Novel Redistribution-based Feature Selection for Text Classification. Expert Syst. Appl. 2024, 246, 123119. [Google Scholar] [CrossRef]
  40. Lazhar, F.; Amira, B. Semantic Similarity-aware Feature selection and Redundancy Removal for Text Classification using joint mutual information. Knowl. Inf. Syst. 2024, 66, 6187–6212. [Google Scholar] [CrossRef]
  41. Rehman, S.U.; Jat, R.; Rafi, M.; Frnda, J. TransFINN “Transparent Feature Integrated Neural Network for Text Feature Selection and Classification”. IEEE Access 2025, 13, 118821–118829. [Google Scholar] [CrossRef]
  42. Chantar, H.; Mafarja, M.; Alsawalqah, H.; Heidari, A.A.; Aljarah, I.; Faris, H. Feature Selection Using Binary Grey Wolf Optimizer with Elite-based Crossover for Arabic Text Classification. Neural Comput. Appl. 2020, 32, 12201–12220. [Google Scholar] [CrossRef]
  43. Alsaleh, D.; Larabi-Marie-Sainte, S. Arabic Text Classification Using Convolutional Neural Network and Genetic Algorithms. IEEE Access 2021, 9, 91670–91685. [Google Scholar] [CrossRef]
  44. Thirumoorthy, K.; Muneeswaran, K. Feature Selection Using Hybrid Poor and Rich Optimization Algorithm for Text Classification. Pattern Recognit. Lett. 2021, 147, 63–70. [Google Scholar] [CrossRef]
  45. Hosseinalipour, A.; Gharehchopogh, F.S.; Masdari, M.; Khademi, A. A Novel Binary Farmland Fertility Algorithm for Feature Selection in Analysis of the Text Psychology. Appl. Intell. 2021, 51, 4824–4859. [Google Scholar] [CrossRef]
  46. Adel, A.; Omar, N.; Abdullah, S.; Al-Shabi, A. Co-Operative Binary Bat Optimizer with Rough Set Reducts for Text Feature Selection. Appl. Sci. 2022, 12, 1296. [Google Scholar] [CrossRef]
  47. Priya, S.K.; Karthika, K.P. An Embedded Feature Selection Approach for Depression Classification Using Short Text Sequences. Appl. Soft Comput. 2023, 147, 110828. [Google Scholar] [CrossRef]
  48. Singh, G.; Nagpal, A.; Singh, V. Optimal Feature Selection and Invasive Weed Tunicate Swarm Algorithm-based Hierarchical Attention Network for Text Classification. Connect. Sci. 2023, 35, 2231171. [Google Scholar] [CrossRef]
  49. Wu, X.; Fei, M.R.; Wu, D.K. Enhanced Binary Black Hole algorithm for text feature selection on resources classification. Knowl.-Based Syst. 2023, 274, 110635. [Google Scholar] [CrossRef]
  50. Kaya, C.; Kilimci, Z.H.; Uysal, M.; Kaya, M. Migrating Birds Optimization-based Feature Selection for Text Classification Fi cation. PeerJ Comput. Sci. 2024, 10, e2263. [Google Scholar] [CrossRef]
  51. Msallam, M.M.; Bin Idris, S.A. Unsupervised Text Feature Selection by Binary Fire Hawk Optimizer for Text Clustering. Clust. Comput.—J. Netw. Softw. Tools Appl. 2024, 27, 7721–7740. [Google Scholar] [CrossRef]
  52. Nachaoui, M.; Lakouam, I.; Hafidi, I. Hybrid Particle Swarm Optimization Algorithm for Text Feature Selection Problems. Neural Comput. Appl. 2024, 36, 7471–7489. [Google Scholar] [CrossRef]
  53. Dhal, P.; Azad, C. A Fine-tuning Deep Learning with Multi-objective-based Feature Selection Approach for the Classification of Text. Neural Comput. Appl. 2024, 36, 3525–3553. [Google Scholar] [CrossRef]
  54. Farek, L.; Benaidja, A. An Adaptive Binary Particle Swarm Optimization Algorithm with Filtration and Local Search for Feature Selection in Text Classification. Memetic Comput. 2025, 17, 45. [Google Scholar] [CrossRef]
  55. El-Hajj, W.; Hajj, H. An Optimal Approach for Text Feature Selection. Comput. Speech Lang. 2022, 74, 101364. [Google Scholar] [CrossRef]
  56. Saeed, M.M.; Al Aghbari, Z. ARTC: Feature Selection Using Association Rules for Text Classification. Neural Comput. Appl. 2022, 34, 22519–22529. [Google Scholar] [CrossRef]
  57. Farghaly, H.M.; Abd El-Hafeez, T. A High-quality Feature Selection Method Based on Frequent and Correlated Items for Text Classification. Soft Comput. 2023, 27, 11259–11274. [Google Scholar] [CrossRef]
  58. Sagbas, E.A. A Novel Two-stage Wrapper Feature Selection Approach Based on Greedy Search for Text Sentiment Classification. Neurocomputing 2024, 590, 127729. [Google Scholar] [CrossRef]
  59. Liu, Y.; Cheng, X.; Stephen, L.S.; Wei, S.S. Advancing Text Classification: A Novel Two-stage Multi-objective Feature Selection Framework. Inf. Technol. Manag. 2025. [Google Scholar] [CrossRef]
  60. Jalilian, E.; Hofbauer, H.; Uhl, A. Iris Image Compression Using Deep Convolutional Neural Networks. Sensors 2022, 22, 2698. [Google Scholar] [CrossRef]
  61. Xiao, Q.G.; Li, G.Y.; Chen, Q.C. Complex image classification by feature inference. Neurocomputing 2023, 544, 126231. [Google Scholar] [CrossRef]
  62. Tyagi, S.; Yadav, D. ForensicNet: Modern convolutional neural network-based image forgery detection network. J. Forensic Sci. 2023, 68, 461–469. [Google Scholar] [CrossRef]
  63. Wang, J.R.; Li, J.; Zhang, Y.R. Text3D: 3D Convolutional Neural Networks for Text Classification. Electronics 2023, 12, 3087. [Google Scholar] [CrossRef]
  64. Xu, S.J.; Guo, C.Y.; Zhu, Y.H.; Liu, G.G.; Xiong, N.L. CNN-VAE: An Intelligent Text Representation Algorithm. J. Supercomput. 2023, 79, 12266–12291. [Google Scholar] [CrossRef]
  65. Gasmi, K.; Ayadi, H.; Torjmen, M. Enhancing Medical Image Retrieval with UMLS-Integrated CNN-Based Text Indexing. Diagnostics 2024, 14, 1204. [Google Scholar] [CrossRef] [PubMed]
  66. Liu, X.Z.; Wang, Y.Y.; Niu, N.N.; Zhang, B.Y.; Li, J.S. A Hybrid Architecture for Enhancing Chinese Text Processing Using CNN and LLaMA2. Sci. Rep. 2025, 15, 24735. [Google Scholar] [CrossRef] [PubMed]
  67. Wahyu Trisna, K.; Huang, J.J.; Chen, Y.J.; Putra, I.G.J.E. Dynamic Text Augmentation for Robust Sentiment Analysis: Enhancing Model Performance with EDA and Multi-Channel CNN. IEEE Access 2025, 13, 31978–31991. [Google Scholar] [CrossRef]
  68. Wu, X.; Cai, Y.; Li, Q.; Xu, J.Y.; Leung, H.F. Combining Weighted Category-aware Contextual Information in Convolutional Neural Networks for Text Classification. World Wide Web-Internet Web Inf. Syst. 2020, 23, 2815–2834. [Google Scholar] [CrossRef]
  69. Xu, J.Y.; Cai, Y.; Wu, X.; Lei, X.; Huang, Q.B.; Leung, H.F.; Li, Q. Incorporating Context-relevant Concepts into Convolutional Neural Networks for Short Text Classification. Neurocomputing 2020, 386, 42–53. [Google Scholar] [CrossRef]
  70. Butt, M.A.; Khatak, A.M.; Shafique, S.; Hayat, B.; Abid, S.; Kim, K.I.; Ayub, M.W.; Sajid, A.; Adnan, A. Convolutional Neural Network Based Vehicle Classification in Adverse Illuminous Conditions for Intelligent Transportation Systems. Complexity 2021, 2021, 6644861. [Google Scholar] [CrossRef]
  71. Liang, Y.J.; Li, H.H.; Guo, B.; Yu, Z.W.; Zheng, X.L.; Samtani, S.; Zeng, D.D. Fusion of Heterogeneous Attention Mechanisms in Multi-view Convolutional Neural Network for Text Classification. Inf. Sci. 2021, 548, 295–312. [Google Scholar] [CrossRef]
  72. Liu, X.Y.; Tang, T.; Ding, N. Social Network Sentiment Classification Method Combined Chinese Text Syntax with Graph Convolutional Neural Network. Egypt. Inform. J. 2022, 23, 1–12. [Google Scholar] [CrossRef]
  73. Zhao, W.D.; Zhu, L.; Wang, M.; Zhang, X.L.; Zhang, J.M. WTL-CNN: A News Text Classification Method of Convolutional Neural Network Based on Weighted Word Embedding. Connect. Sci. 2022, 34, 2291–2312. [Google Scholar] [CrossRef]
  74. Qorich, M.; El Ouazzani, R. Text Sentiment Classification of Amazon Reviews Using Word Embeddings and Convolutional Neural Networks. J. Supercomput. 2023, 79, 11029–11054. [Google Scholar] [CrossRef]
  75. Thekkekara, J.P.; Yongchareon, S.; Liesaputra, V. An Attention-based CNN-BiLSTM Model for Depression Detection on Social Media Text. Expert Syst. Appl. 2024, 249, 123834. [Google Scholar] [CrossRef]
  76. Guo, X.L.; Liu, Q.Y.; Hu, Y.R.; Liu, H.J. MDCNN: Multi-Teacher Distillation-Based CNN for News Text Classification. IEEE Access 2025, 13, 56631–56641. [Google Scholar] [CrossRef]
  77. Liu, B.; Zhou, Y.; Sun, W. Character-level Text Classification via Convolutional Neural Network and Gated Recurrent Unit. Int. J. Mach. Learn. Cybern. 2020, 11, 1939–1949. [Google Scholar] [CrossRef]
  78. Zeng, S.F.; Ma, Y.; Zhang, X.Y.; Du, X.F. Term-Based Pooling in Convolutional Neural Networks for Text Classification. China Commun. 2020, 17, 109–124. [Google Scholar] [CrossRef]
  79. Lyu, S.F.; Liu, J.Q. Convolutional Recurrent Neural Networks for Text Classification. J. Databased Manag. 2021, 32, 65–82. [Google Scholar] [CrossRef]
  80. Wang, H.T.; Tian, K.K.; Wu, Z.J.; Wang, L. A Short Text Classification Method Based on Convolutional Neural Network and Semantic Extension. Int. J. Comput. Intell. Syst. 2021, 14, 367–375. [Google Scholar] [CrossRef]
  81. Liu, J.F.; Ma, H.Z.; Xie, X.L.; Cheng, J. Short Text Classification for Faults Information of Secondary Equipment Based on Convolutional Neural Networks. Energies 2022, 15, 2400. [Google Scholar] [CrossRef]
  82. Zhou, Y.J.; Li, J.L.; Chi, J.H.; Tang, W.; Zheng, Y.Q. Set-CNN: A Text Convolutional Neural Network Based on Semantic Extension for Short Text Classification. Knowl.-Based Syst. 2022, 257, 109948. [Google Scholar] [CrossRef]
  83. Xiong, Y.P.; Chen, G.L.; Cao, J.K. Research on Public Service Request Text Classification Based on BERT-BiLSTM-CNN Feature Fusion. Appl. Sci. 2024, 14, 6282. [Google Scholar] [CrossRef]
  84. Li, J.H.; Yang, Y.; Sun, J.; Wang, F.; Chen, S.L. DT-GCNN: Dynamic Triplet Network with GRU-CNN for Enhanced Text Classification. Int. J. Mach. Learn. Cybern. 2025, 16, 9555–9567. [Google Scholar] [CrossRef]
  85. Huang, M.H.; Xie, H.R.; Rao, Y.H.; Feng, J.R.; Wang, F.L. Sentiment Strength Detection with a Context-dependent Lexicon-based Convolutional Neural Network. Inf. Sci. 2020, 520, 389–399. [Google Scholar] [CrossRef]
  86. Krishnan, H.; Elayidom, M.S.; Santhanakrishnan, T. Optimization Assisted Convolutional Neural Network for Sentiment Analysis with Weighted Holoentropy-based Features. Int. J. Inf. Technol. Decis. Mak. 2021, 20, 1261–1297. [Google Scholar] [CrossRef]
  87. Usama, M.; Ahmad, B.; Song, E.; Hossain, M.S.; Alrashoud, M.; Muhammad, G. Attention-based Sentiment Analysis Using Convolutional and Recurrent Neural Network. Future Gener. Comput. Syst. 2020, 113, 571–578. [Google Scholar] [CrossRef]
  88. Wang, X.Y.; Li, F.; Zhang, Z.Q.; Xu, G.L.; Zhang, J.Y.; Sun, X. A Unified Position-aware Convolutional Neural Network for Aspect Based Sentiment Analysis. Neurocomputing 2021, 450, 91–103. [Google Scholar] [CrossRef]
  89. Ghorbanali, A.; Sohrabi, M.K.; Yaghmaee, F. Ensemble Transfer Learning-based Multimodal Sentiment Analysis Using Weighted Convolutional Neural Networks. Inf. Process. Manag. 2022, 59, 102929. [Google Scholar] [CrossRef]
  90. Huang, M.H.; Xie, H.R.; Rao, Y.H.; Liu, Y.W.; Poon, L.K.M.; Wang, F.L. Lexicon-Based Sentiment Convolutional Neural Networks for Online Review Analysis. IEEE Trans. Affect. Comput. 2022, 13, 1337–1348. [Google Scholar] [CrossRef]
  91. Murugaiyan, S.; Uyyala, S.R. Aspect-Based Sentiment Analysis of Customer Speech Data Using Deep Convolutional Neural Network and BiLSTM. Cogn. Comput. 2023, 15, 914–931. [Google Scholar] [CrossRef]
  92. Mutinda, J.; Mwangi, W.; Okeyo, G. Sentiment Analysis of Text Reviews Using Lexicon-Enhanced Bert Embedding (LeBERT) Model with Convolutional Neural Network. Appl. Sci. 2023, 13, 1445. [Google Scholar] [CrossRef]
  93. Alnowaiser, K. Scientific Text Citation Analysis Using CNN Features and Ensemble Learning Model. PLoS ONE 2024, 19, e0302304. [Google Scholar] [CrossRef]
  94. He, A.X.; Abisado, M. Text Sentiment Analysis of Douban Film Short Comments Based on BERT-CNN-BiLSTM-Att Model. IEEE Access 2024, 12, 45229–45237. [Google Scholar] [CrossRef]
  95. Chen, J.Y.; Yan, S.K.; Wong, K.C. Verbal Aggression Detection on Twitter Comments: Convolutional Neural Network for Short-text Sentiment Analysis. Neural Comput. Appl. 2020, 32, 10809–10818. [Google Scholar] [CrossRef]
  96. Qiu, M.; Zhang, Y.R.; Ma, T.Q.; Wu, Q.F.; Jin, F.Z. Convolutional-neural-network-based Multilabel Text Classification for Automatic Discrimination of Legal Documents. Sens. Mater. 2020, 32, 2659–2672. [Google Scholar] [CrossRef]
  97. Heo, T.S.; Kim, C.; Kim, J.D.; Park, C.Y.; Kim, Y.S. Prediction of Atrial Fibrillation Cases: Convolutional Neural Networks Using the Output Texts of Electrocardiography. Sens. Mater. 2021, 33, 393–404. [Google Scholar] [CrossRef]
  98. Li, Y.; Yin, C.B. Application of Dual-Channel Convolutional Neural Network Algorithm in Semantic Feature Analysis of English Text Big Data. Comput. Intell. Neurosci. 2021, 2021, 7085412. [Google Scholar] [CrossRef]
  99. Jian, L.H.; Xiang, H.Q.; Le, G.B. English Text Readability Measurement Based on Convolutional Neural Network: A Hybrid Network Model. Comput. Intell. Neurosci. 2022, 2022, 6984586. [Google Scholar] [CrossRef]
  100. Qiu, Q.J.; Xie, Z.; Ma, K.; Chen, Z.L.; Tao, L.F. Spatially Oriented Convolutional Neural Network for Spatial Relation Extraction from Natural Language Texts. Trans. GIS 2022, 26, 839–866. [Google Scholar] [CrossRef]
  101. Boukhers, Z.; Goswami, P.; Jurjens, J. Knowledge Guided Multi-filter Residual Convolutional Neural Network for ICD Coding from Clinical Text. Neural Comput. Appl. 2023, 35, 17633–17644. [Google Scholar] [CrossRef]
  102. Muppudathi, S.S.; Krishnasamy, V. Anomaly Detection in Social Media Texts Using Optimal Convolutional Neural Network. Intell. Autom. Soft Comput. 2023, 36, 1027–1042. [Google Scholar] [CrossRef]
  103. Fan, W.T. BiLSTM-Attention-CNN Model Based on ISSA Optimization for Cyberbullying Detection in Chinese Text. Inf. Technol. Control. 2024, 53, 659–674. [Google Scholar] [CrossRef]
  104. Zeng, Q.X. Enhanced Analysis of Large-scale News Text Data Using the Bidirectional-Kmeans-LSTM-CNN Model. Peerj Comput. Sci. 2024, 10, e2213. [Google Scholar] [CrossRef] [PubMed]
  105. Faseeh, M.; Khan, M.A.; Iqbal, N.; Qayyum, F.; Mehmood, A.; Kim, J. Enhancing User Experience on Q&A Platforms: Measuring Text Similarity Based on Hybrid CNN-LSTM Model for Efficient Duplicate Question Detection. IEEE Access 2024, 12, 34512–34526. [Google Scholar]
  106. Wu, H.C.; Li, H.X.; Fang, X.C.; Tang, M.Q.; Yu, H.Z.; Yu, B.; Li, M.; Jing, Z.R.; Meng, Y.H.; Chen, W.; et al. MDCNN: A Multimodal Dual-CNN Recursive Model for Fake News Detection via Audio- and Text-based Speech Emotion Recognition. Speech Commun. 2025, 175, 103313. [Google Scholar] [CrossRef]
  107. Hong, M.; Wang, H.Y. Filter feature selection methods for text classification: A review. Multimed. Tools Appl. 2024, 83, 2053–2091. [Google Scholar]
  108. Ganesan, K.; Zhai, C.X. Opinion-based Entity Ranking. Inf. Retr. 2012, 15, 116–150. [Google Scholar] [CrossRef]
  109. Ciarelli, P.M.; Oliveira, E. Agglomeration and Elimination of Terms for Dimensionality Reduction. In Proceedings of the 2009 Ninth International Conference on Intelligent Systems Design and Applications, Pisa, Italy, 30 November–2 December 2009; pp. 547–552. [Google Scholar]
  110. Ciarelli, P.M.; Salles, E.O.T.; Oliveira, E. An Evolving System Based on Probabilistic Neural Network. In Proceedings of the 2010 Eleventh Brazilian Symposium on Neural Networks, Sao Paulo, Brazil, 23–28 October 2010; Volume 1, pp. 182–187. [Google Scholar]
  111. Kotzias, D.; Denil, M.; De Freitas, N.; Smyth, P. From Group to Individual Labels Using Deep Features. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, 10–13 August 2015; pp. 597–606. [Google Scholar]
  112. Mustafa, A.M.; Rashid, T.A. Kurdish Stemmer Pre-processing Steps for Improving Information Retrieval. J. Inf. Sci. 2018, 44, 15–27. [Google Scholar] [CrossRef]
  113. Rashid, T.A.; Mustafa, A.M.; Saeed, A.M. A Robust Categorization System for Kurdish Sorani Text Documents. Inf. Technol. J. 2017, 16, 27–34. [Google Scholar] [CrossRef]
  114. Rashid, T.A.; Mustafa, A.M.; Saeed, A.M. Automatic Kurdish Text Classification Using KDC 4007 Dataset. In Proceedings of the 5th International Conference on Emerging Internetworking, Data & Web Technologies, Wuhan, China, 10–11 June 2017; Volume 6, pp. 187–198. [Google Scholar]
  115. Kilinç, D.; Özçift, A.; Bozyiğit, F.; Yildirim, P.; Yucalar, F.; Borandağ, E. Ttc-3600: A New Benchmark Dataset for Turkish Text Categorization. J. Inf. Sci. 2015, 43, 174–185. [Google Scholar] [CrossRef]
  116. Wang, G.; Lochovsky, F.H. Feature Selection with Conditional Mutual Information MaxiMin in Text Categorization. In Proceedings of the 2004 ACM CIKM International Conference on Information and Knowledge Management, Washington, DC, USA, 8–13 November 2004; pp. 342–349. [Google Scholar]
  117. Gao, Z.; Xu, Y.; Meng, F.; Qi, F.; Lin, Z.Q. Improved Information Gain-Based Feature Selection for Text Categorization. In Proceedings of the International Conference on Wireless Communications, Vehicular Technology, Information Theory and Aerospace & Electronic Systems (VITAE), Aalborg, Denmark, 11–14 May 2014; pp. 1–5. [Google Scholar]
  118. Zheng, Z.; Wu, X.; Srihari, R. Feature Selection for Text Categorization on Imbalanced Data. ACM SIGKDD Explor. Newsl. 2004, 6, 80–89. [Google Scholar] [CrossRef]
  119. Shang, W.; Huang, H.; Zhu, H.; Lin, Y.; Qu, Y.; Wang, Z. A Novel Feature Selection Algorithm for Text Categorization. Expert Syst. Appl. 2007, 33, 1–5. [Google Scholar] [CrossRef]
  120. Uğuz, H. A Two-stage Feature Selection Method for Text Categorization by Using Information Gain, Principal Component Analysis and Genetic Algorithm. Knowl.-Based Syst. 2011, 24, 1024–1032. [Google Scholar] [CrossRef]
  121. Azam, N.; Yao, J.T. Comparison of Term Frequency and Document Frequency Based Feature Selection Metrics in Text Categorization. Expert Syst. Appl. 2012, 39, 4760–4768. [Google Scholar] [CrossRef]
  122. Rehman, A.; Javed, K.; Babri, H.A. Feature Selection Based on a Normalized Difference Measure for Text Classification. Inf. Process. Manag. 2017, 53, 473–489. [Google Scholar] [CrossRef]
Figure 1. Structure of the three-layer CNN.
Figure 1. Structure of the three-layer CNN.
Electronics 14 04615 g001
Figure 2. Structure of the smoothing layer.
Figure 2. Structure of the smoothing layer.
Electronics 14 04615 g002
Figure 3. Structure of the extending layer.
Figure 3. Structure of the extending layer.
Electronics 14 04615 g003
Figure 4. Structure of the compressing layer.
Figure 4. Structure of the compressing layer.
Electronics 14 04615 g004
Figure 5. The flow chart of feature selection using CNNFS.
Figure 5. The flow chart of feature selection using CNNFS.
Electronics 14 04615 g005
Figure 6. Classification accuracy of the CF dataset.
Figure 6. Classification accuracy of the CF dataset.
Electronics 14 04615 g006
Figure 7. Classification accuracy of the CR dataset.
Figure 7. Classification accuracy of the CR dataset.
Electronics 14 04615 g007
Figure 8. Classification accuracy of the CNAE dataset.
Figure 8. Classification accuracy of the CNAE dataset.
Electronics 14 04615 g008
Figure 9. Classification accuracy of the IMDB dataset.
Figure 9. Classification accuracy of the IMDB dataset.
Electronics 14 04615 g009
Figure 10. Classification accuracy of the KDC dataset.
Figure 10. Classification accuracy of the KDC dataset.
Electronics 14 04615 g010
Figure 11. Classification accuracy of the TTC dataset.
Figure 11. Classification accuracy of the TTC dataset.
Electronics 14 04615 g011
Figure 12. Classification accuracy of the WEKBE dataset.
Figure 12. Classification accuracy of the WEKBE dataset.
Electronics 14 04615 g012
Figure 13. Classification accuracy of the R8 dataset.
Figure 13. Classification accuracy of the R8 dataset.
Electronics 14 04615 g013
Figure 14. Classification accuracy of the R52 dataset.
Figure 14. Classification accuracy of the R52 dataset.
Electronics 14 04615 g014
Figure 15. Terms related to topics of datasets in the top 20 terms.
Figure 15. Terms related to topics of datasets in the top 20 terms.
Electronics 14 04615 g015aElectronics 14 04615 g015b
Figure 16. Sparsity of document–term matrices.
Figure 16. Sparsity of document–term matrices.
Electronics 14 04615 g016
Table 1. Details of the CNN architecture and training parameters.
Table 1. Details of the CNN architecture and training parameters.
Smoothing LayerExtending LayerCompression Layer
LayerParamsLayerParamsLayerParams
A convolution layer 3 × 3 kernel
1 × 1 padding
Four smoothing layersA convolution layer 3 × 3 kernel
1 × 1 padding
A maxpooling layer3 × 3 kernel
1 × 1 padding
A maxpooling layer2 × 2 kernel
2 × 2 strides
A fully connected layer D × | D | weight matrix
Table 2. Details of the nine datasets.
Table 2. Details of the nine datasets.
Name of DatasetQuantity of DocumentsQuantity of TermsNumber
of Classes
Description
CF1007182412User favorite description text of 12 different Toyota cars
CR1047395612User review text of 12 different Toyota cars
CNAE10808569Business description text of Brazilian companies from nine groups
IMDB100024222Positive and negative movie review text
KDC400713,2018Turkish text of articles and news
TTC360032086
WEBKB419976844Collection of web pages of computer science departments of four universities
R8767417,2318Subset generated from the Reuters-21578 dataset
R52910019,08052
Table 3. Comparative feature selection methods.
Table 3. Comparative feature selection methods.
MethodDescription
Information Gain (IG) I G t j = P t j c k C L P c k | t j log P c k | t j P c k + P t ¯ j c k C L P c k | t ¯ j log P c k | t ¯ j P c k
Expected Cross Entropy (ECE) E C E t j = P t j c k C L P c k | t j log P c k | t j P c k
Mutual Information (MI) M I t j   = c k C L P ( c k ) l o g P c k   | t j P c k
Gini Index (GI) G I t j = c k C L P c k | t j 2
Weighted Odd Ratio (WOR) W O R t j = c k C L P c k l o g P t j | c k [ 1 P t j | c ¯ k ] [ 1 P t j | c k ] P t j | c ¯ k
Normalized Difference Measure (NDM) N D M t j = c k C L P ( c k ) P t j c k P t j c ¯ k m i n ( P t j c k , P t j c ¯ k )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, J.; Hong, M. A Feature Selection Method Based on a Convolutional Neural Network for Text Classification. Electronics 2025, 14, 4615. https://doi.org/10.3390/electronics14234615

AMA Style

Xiao J, Hong M. A Feature Selection Method Based on a Convolutional Neural Network for Text Classification. Electronics. 2025; 14(23):4615. https://doi.org/10.3390/electronics14234615

Chicago/Turabian Style

Xiao, Jiongen, and Ming Hong. 2025. "A Feature Selection Method Based on a Convolutional Neural Network for Text Classification" Electronics 14, no. 23: 4615. https://doi.org/10.3390/electronics14234615

APA Style

Xiao, J., & Hong, M. (2025). A Feature Selection Method Based on a Convolutional Neural Network for Text Classification. Electronics, 14(23), 4615. https://doi.org/10.3390/electronics14234615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop