Next Article in Journal
Groundwater Detection Using the Pseudo-3D Resistivity Method: A History of Case Studies
Next Article in Special Issue
Neural Networks for Early Diagnosis of Postpartum PTSD in Women after Cesarean Section
Previous Article in Journal
Revisiting the Decomposition Process of Tetrahydrate Co(II) Acetate: A Sample’s Journey through Temperature
Previous Article in Special Issue
A High-Level Representation of the Navigation Behavior of Website Visitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Chicken Swarm-Based Feature Subset Selection with Optimal Machine Learning Enabled Data Mining Approach

by
Monia Hamdi
1,
Inès Hilali-Jaghdam
2,
Manal M. Khayyat
3,
Bushra M. E. Elnaim
4,
Sayed Abdel-Khalek
5,6 and
Romany F. Mansour
7,*
1
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Computer Sciences and Information Technology, Applied College, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
3
Department of Information Systems, College of Computers and Information Systems, Umm Al-Qura University, P.O. Box 7607, Makkah 24382, Saudi Arabia
4
Department of Computer Science, College of Science and Humanities in Al-Sulail, Prince Sattam Bin Abdulaziz University, P.O. Box 173, Al-Kharj 16278, Saudi Arabia
5
Department of Mathematics, Faculty of Science, Sohag University, Sohag 82524, Egypt
6
Department of Mathematics, College of Science, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
7
Department of Mathematics, Faculty of Science, New Valley University, El-Kharga 72511, Egypt
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6787; https://doi.org/10.3390/app12136787
Submission received: 9 April 2022 / Revised: 18 June 2022 / Accepted: 29 June 2022 / Published: 4 July 2022
(This article belongs to the Special Issue Data Analysis and Mining)

Abstract

:
Data mining (DM) involves the process of identifying patterns, correlation, and anomalies existing in massive datasets. The applicability of DM includes several areas such as education, healthcare, business, and finance. Educational Data Mining (EDM) is an interdisciplinary domain which focuses on the applicability of DM, machine learning (ML), and statistical approaches for pattern recognition in massive quantities of educational data. This type of data suffers from the curse of dimensionality problems. Thus, feature selection (FS) approaches become essential. This study designs a Feature Subset Selection with an optimal machine learning model for Educational Data Mining (FSSML-EDM). The proposed method involves three major processes. At the initial stage, the presented FSSML-EDM model uses the Chicken Swarm Optimization-based Feature Selection (CSO-FS) technique for electing feature subsets. Next, an extreme learning machine (ELM) classifier is employed for the classification of educational data. Finally, the Artificial Hummingbird (AHB) algorithm is utilized for adjusting the parameters involved in the ELM model. The performance study revealed that FSSML-EDM model achieves better results compared with other models under several dimensions.

1. Introduction

Data mining (DM) is the procedure of understanding data through cleaning raw data, discovering patterns, producing models, and testing the models. It comprises of several fields such as statistics, machine learning (ML), and database systems. Education DM (EDM) is an emergent field with an arising strategy to investigate the various types of data which are obtained from an education background [1]. It is an interdisciplinary area which inspects data mining (DM), man-fabricated consciousness, and measurable demonstrating with the data produced using an academic organization [2]. EDM uses a calculation method for taking care of elucidating academic data considering a definitive point of examining academic enquiries. To make a nation stand out among different nations across the globe, education frameworks should encounter an essential advancement by re-planning their design. The concealed data and examples from various data sources can be extricated by adjusting the strategies for DM. For summing up the outcomes of students with their qualifications, they investigate the abuse of DM in the academic fields. Crude data can be altogether moved through DM models. The data achieved from an education organization go through examination of various DM strategies [3]. The strategy identifies the conditions wherein students can strive to have a positive impact [4].
Student performance prediction (SPP) has different definitions according to troublesome perspectives; however, the measured assessment assumes a significant part in current education establishments. SPP seems effective in aiding all partners in the education interaction. For students, SPP can assist them with picking reasonable courses or activities and make their arrangements for academic durations [5]. For educators, SPP can assist with changing learning material and presenting programs compatible with the students’ capacity, and help identify struggling students. For education chiefs, SPP can assist with checking the education program and enhancing the course framework. Generally, partners in the education advancement have intentions to further develop the education outcome. Moreover, the data-driven SPP study provides a goal of reference for the education framework. Weka, a compelling DM technique was utilized to produce the outcome [6].
The increment of educational information from distinct sources has resulted in desperation for the EDM research [7]. This can help to further objectives and characterize specific goals of education and highlight subset determinations by disposing of the component that is repetitive/is not important. The set of components chosen should follow the Occam Razor rule to provide the best outcome in light of the goal [8]. The data size to be dealt with has expanded in the past five years; hence, the choice is turning into a necessity before any sort of arrangement happens. It is not quite the same as the element extraction strategy because the determination method does not change the first portrayal of the data [9]. The least complex method includes choice, where how much quality in an examination is diminished by choosing just the main view of the circumstances such as a more elevated level of exercises [10].
Since the high dimensionality raises computational costs, it is essential to define a way to reduce the number of considered features. Feature selection (FS) allows reducing a high dimensionality problem and selecting a suitable number of features. This study designed a Feature Subset Selection with optimal machine learning for the Educational Data Mining (FSSML-EDM) model. The proposed FSSML-EDM model involves the Chicken Swarm Optimization-based Feature Selection (CSO-FS) technique for electing feature subsets. Next, the extreme learning machine (ELM) classifier was employed for the classification of educational data. Finally, the Artificial Hummingbird (AHB) algorithm was utilized for adjusting the parameters involved in the ELM model. The performance study revealed the effectual outcomes of the FSSML-EDM model over the compared models under several dimensions. Our contributions are summarized as follows:
  • We propose a model comprising data preprocessing, CSO-FS, ELM classification, and AHB parameter;
  • We designed a new CSO-FS technique to reduce the curse of the dimensionality problem and enhanced the classification performance;
  • We employed the ELM classification model with the AHB-based parameter optimization technique for the EDM process;
  • We validated the performance of the FSSML-EDM model using the benchmark dataset from the UCI repository.

2. Literature Review

Injadat et al. [11] explored and analyzed two distinct datasets at two distinct phases of course delivery (20% and 50%) utilizing several graphical, statistical, and quantitative approaches. The feature analysis offers understanding as to the nature of distinct features regarded and utilizes in the selection of ML techniques and their parameters. Moreover, this work presents a systematic model dependent upon the Gini index and p-value for selecting an appropriate ensemble learner in a group of six potential ML techniques. Ashraf et al. [12] progressed an accurate prediction pedagogical method, considering the pronounced nature and novelty of presented approach in Educational Data Mining. The base classifications containing RT, j48, kNN, and naïve Bayes (NB) were estimated on a 10-fold cross-validation model. In addition, the filter procedure as over-sampling (SMOTE) and under-sampling (Spread subsampling) were exploited to examine some important alterations in outcomes amongst meta and base classifications.
Dabhade et al. [13] forecasted student academic performance in a technical institution in India. The data pre-processed and factor analysis were executed on the attained dataset for removing the anomaly from the data, decreasing the dimensionality of the data, and attaining the most correlated feature. Nahar et al. [14] generated two datasets concentrating on two distinct angles. In the primary dataset classification and forecast, the type of students (bad, medium, and good) on a particular course was dependent upon its prerequisite course efficiency. This can be executed during the artificial intelligence (AI) course. The secondary dataset also classified and forecasted the last grade (A, B, C) of an arbitrary subject; our data can be established in such a way that the data are only concentrated on the efficiency of the midterm exam.
Despite all the studies performed on FS process, to the best of our knowledge, only few works have carried out an FS-based classification model for EDM. Earlier works have used ML models for EDM without contributing much significance to the FS process. At the same time, the parameters involved in the ML models (i.e., ELM) considerably affect the overall classification performance. Since the trial-and-error method for parameter tuning is a tedious and erroneous process, metaheuristic algorithms can be applied. Therefore, metaheuristic optimization algorithms can be designed to optimally tune the parameters related to the ML models to improve the overall classification performance.

3. The Proposed Model

In this study, a new FSSML-EDM technique was developed for mining educational data. The proposed FSSML-EDM model involves data preprocessing at the initial stage to transform the input data into a compatible format. Then, the preprocessed data are passed into the CSO-FS technique for electing feature subsets. Next, the ELM classifier can be employed for the effective identification and classification of educational data. Finally, the AHB algorithm is utilized for effectively adjusting the parameters involved in the ELM model. The outcome of the ELM model is the classification output. Figure 1 depicts the block diagram of FSSML-EDM technique.

3.1. Process Involved in CSO-FS Technique

At the initial stage, the presented FSSML-EDM model incorporates the design of the CSO-FS technique for electing feature subsets. The CSO algorithm is chosen over other optimization algorithms due to its simplicity and high parallelism. The CSO simulates the chicken movement and the performance of the chicken swarm; the CSO is explained as follows: CSO has several groups, and all the groups have a dominant rooster, some hens, and chicks [15]. The rooster, hen, and chick from the group are found dependent upon their fitness value. The rooster (group head) is the chicken which is an optimum fitness value. However, the chick is the chicken which has the worse fitness value. The majority of chickens are hens and it can be selected arbitrarily to stay in that group. The dominance connection and mother–child connection from the group remain unaltered and upgrade during (G) time steps. The movement of chickens are expressed under the equation which utilizes to the rooster place upgrade provided by Equation (1):
X i j r + 1 = X i , j t × ( 1 + r a n d n ( 0 ,   σ 2 ) )
whereas:
σ 2 = { 1 if   f i f k exp ( f k f i | f i + ε | ) Otherwise
In which k [ 1 ,   N r ] , k i , and Nr refers the amount of chosen roosters. X i , j signifies the place of rooster number i in j th dimensional under t and t + 1 iteration, r a n d n   ( O ,   σ 2 ) is utilized for generating Gaussian arbitrary numbers with mean 0 and variance σ 2 ; ε refers to the constant with minimum value; and f i is the fitness value to the equivalent rooster i . The equation which utilizes the hen place upgrade is provided by Equations (2)–(4):
X i , j t + 1 = X i , j t + S 1 r a n d n ( X r 1 , j t X i , j t ) + S 2 r a n d n ( X r 2 , j t X i , j t )
In which:
S 1 = exp ( f i f r 1 | f i | + ε )
and:
S 1 =   exp   ( f r 2 f i )
where r 1 , r [ 1 ,   ,   N ] ,   r 1 r r refers to the index of the rooster, but r 2 implies the chicken in the swarm which is a rooster or hen and a uniform arbitrary number is created by r a n d n . Finally, the equation that utilizes the chick place upgrade is provided by Equation (5):
X i , j t + 1 = X i , j t + F L ( X m , j t X i , j t ) ,   F L [ 0 , 2 ]
where X m , j t signifies the place of i th chick mother.

3.2. ELM Based Classification

At this stage, the ELM classifier can be employed for the effective identification and classification of educational data. The ELM model has n input layers, l hidden layers, and m output layers. Initially, considering the training instance { X ,   Y } = { x i ,   y i } ( i = 1 , 2 ,   ,   Q ) , and it is comprised of the input feature X = [ x i 1 x i 2   x i Q ] and matrix Y = [ y j 1 y j 2 y j Q ] with training instance, where the matrix X and Y are expressed by [16]:
X = [ x 11 x 12 x 1 Q x 21 x 22 x 2 Q x n 1 x n 2 x n Q ] ,  
Y = [ y 11 y 12 y m Q y 21 y 22 y m Q y m 1 y m 2 y m Q ] ,
where n and m parameters denote the dimension of input and output matrix. Next, ELM set the weights amongst input and hidden layers randomly:
w = [ w 11 w 12 w 1 n w 21 w 22 w 2 n w l 1 w l 2 w l n ] ,  
where w i j denotes the weight from i th hidden and j th input layers. Figure 2 showcases the framework of ELM. Then, ELM considers the weight from output and hidden layers that can be shown below:
β = [ β 11 β 12 β 1 m β 21 β 22 β 2 m β l 1 β l 2 β l m ] ,
where β j k indicates the weight from j t h hidden and ith output layers. Next, ELM set a bias of hidden layers randomly:
B = [ b 1 b 2     b n ] T .
After that, ELM chooses the network activation function. According to, output matrix T is characterized by:
T = [ t 1 ,   t 2 ,   ,   t Q ] m × Q .
The column vector of output matrix T is shown in the following:
t j = [ t 1 j t 2 j t m j ] = [ i = 1 l β i 1 g ( w i x j + b i ) i = 1 l β i 2 g ( w i x j + b i ) i = 1 l β i m g ( w i x j + b i ) ] ( j = 1 , 2 , 3 ,   ,   Q ) .  
Moreover, considering Equations (10) and (11), it is modelled in the following:
H β = T ,
where T indicates transpose of T and H represents simulation outcomes of hidden neurons. To achieve better solution with less error, the least square model is employed for determining the weight matrix measure of β [17,18].
β = H + T .
To enhance the normalization ability of the system and provide stable outcomes, the regularization parameter of β is used. The amount of hidden layers is minimal in contrast with the amount of training samples, β is shown in the following [19]:
β = ( I λ + H T H ) 1 H T T .
When the amount of hidden layers is maximal than the amount of training samples, β is denoted as follows [20]:
β = H T ( I λ + H H T ) 1 T .

3.3. AHB Based Parameter Optimization

Finally, the AHB algorithm is utilized for effectively adjusting the parameters involved in the ELM model with the goal of attaining maximum classification performance [21]. The AHB algorithm is an optimization approach stimulated from the foraging and flight of hummingbirds. The three major models are provided as follows: in a guided foraging model, three flight behaviors are utilized in foraging (axial, diagonal, and omnidirectional flight). It can be defined as follows:
v i ( t + 1 ) = x i , t a ( t ) + h · b · ( x i ( t ) X i , t a ( t ) ) h N ( 0 , 1 )
where x i , t a ( t ) characterizes the location of the targeted food source, h signifies the guiding factor, and x ( t ) represents the location of i th food source at time t. The location updating of the ith food source is provided by:
x A i ( t ) = { x i ( t ) f ( x i ( t ) ) f ( v i ( t + 1 ) ) v i ( t + 1 ) f ( x i ( t ) ) > f ( v i ( t + 1 ) )  
where f ( x i ( t ) ) and f ( v i ( t + 1 ) ) denote the value of function fitness for x ( t ) and v i ( t + 1 ) . the local search of hummingbirds in the territorial foraging strategy is provided in the following:
v i ( t + 1 ) = x i ( t ) + g · b · ( χ i ( t ) ) g N ( 0 , 1 )
where g represents the territorial factor. The arithmetical formula for the migration foraging of hummingbirds is provided by:
x w o r ( t + 1 ) = l b + r · ( u b l b )  
where x w o r indicates the source of food with worst population rate of nectar refilling, r represents a random factor, and u b and l b denote the upper and lower limits, respectively.

4. Experimental Validation

The proposed FSSML-EDM model was simulated using a benchmark dataset from UCI repository, which comprises of 649 samples with 32 features and 2 class labels as illustrated in Table 1. The parameter settings are provided as follows: learning rate, 0.01; dropout, 0.5; batch size, 5; and number of epochs, 50. For experimental validation, the dataset is split into 70% training (TR) data and 30% testing (TS) data.
Figure 3 highlights a set of confusion matrices produced by the FSSML-EDM model on the test data. The figure indicates that the FSSML-EDM model resulted in effectual outcomes. On entire dataset, the FSSML-EDM model identified 545 samples into pass and 86 samples into fail. In addition, on 70% of the training dataset, the FSSML-EDM model identified 390 samples into pass and 53 samples into fail. Moreover, on 30% of testing dataset, the FSSML-EDM model identified 155 samples into pass and 33 samples into fail.
Table 2 offers a comprehensive EDM outcome of the FSSML-EDM model on test dataset. The experimental values indicated that the FSSML-EDM model accomplished maximum outcomes on all datasets. Figure 4 provides brief classification results of the FSSML-EDM model on entire dataset. It can be inferred from the figure that the FSSML-EDM model classified pass instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 97.23%, 97.50%, 99.27%, 98.38%, and 89.08% respectively. Moreover, the figure shows that the FSSML-EDM model classified fail instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 97.23%, 95.56%, 86%, 90.53%, and 89.08%, respectively.
Figure 5 provides detailed classification results of the FSSML-EDM model on 70% of the training dataset. The figure reveals that the FSSML-EDM model classified pass instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 97.58%, 97.50%, 99.74%, 98.61%, and 89.57%, respectively. In addition, the figure shows that the FSSML-EDM model classified fail instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 97.58%, 98.15%, 84.13%, 90.60%, and 89.57% respectively.
Figure 6 offers brief classification results of the FSSML-EDM approach on 30% of testing dataset. The figure exposes that the FSSML-EDM algorithm classified pass instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 96.41%, 97.48%, 98.1%, 97.798%, and 88.22%, respectively. Moreover, the figure shows that the FSSML-EDM approach classified fail instances for a c c u y ,   p r e c n , r e c a l , F s c o r e ,   M C C , and kappa of 96.41%, 91.67%, 89.19%, 90.41%, and 88.22%, respectively.
Figure 7 illustrates the training and validation accuracy inspection of the FSSML-EDM technique on the applied dataset. The figure of the FSSML-EDM approach offers maximum training/validation accuracy on the classification process.
Next, Figure 8 reveals the training and validation loss inspection of the FSSML-EDM approach on the applied dataset. The figure shows that the FSSML-EDM algorithm offers reduced training/accuracy loss on the classification process of the test data.
A brief precision-recall examination of the FSSML-EDM model on the test dataset is portrayed in Figure 9. By observing the figure, it is noticed that the DLBTDC-MRI model accomplished maximum precision-recall performance under all classes.
A detailed ROC investigation of the FSSML-EDM approach on the distinct datasets is portrayed in Figure 10. The results indicate that the FSSML-EDM technique exhibited its ability in categorizing two different classes such as pass and fail on the test datasets.
Table 3 reveals an extensive comparative study of the FSSML-EDM model with existing models such as improved evolutionary algorithm-based feature subsets election with neuro-fuzzy classification (IEAFSSNFC) [22], neuro-fuzzy classification (NFC) [21], neural network (NN), support vector machines (SVM), decision tree (DT), and random forest (RF). Figure 11 inspects the comparative p r e c n , r e c a l , and a c c u y investigation of the FSSML-EDM model with recent methods. The figure reveals that the NN and SVM models showed poor performance with lower values of p r e c n , r e c a l , and a c c u y . The NFC, DT, and RF models have showed slightly improved values of p r e c n , r e c a l , and a c c u y . Moreover, the IEAFSSNFC model resulted in reasonable p r e c n , r e c a l , and a c c u y of 93.81%, 92.39%, and 90.33%, respectively. Furthermore, the FSSML-EDM model accomplished effectual outcomes with maximum p r e c n , r e c a l , and a c c u y of 94.58%, 93.65%, and 96.41%, respectively.
Figure 12 inspects the comparative F s c o r e , M C C , and k a p p a analysis of the FSSML-EDM method with existing algorithms. The figure reveals that the NN and SVM methods showed poor performance with lower values of F s c o r e , M C C , and k a p p a . Similarly, the NFC, DT, and RF approaches showed slightly improved values of F s c o r e , M C C , and k a p p a .
Moreover, the IEAFSSNFC model resulted a in reasonable F s c o r e , M C C , and k a p p a of 93.01%, 73.78%, and 73.37%, respectively. Additionally, the FSSML-EDM methodology accomplished effectual outcomes with maximum F s c o r e , M C C , and k a p p a of 94.10%, 82.22%, and 88.20%, respectively. Therefore, the FSSML-EDM model has the capability of assessing student performance in real time.

5. Conclusions

In this study, a new FSSML-EDM technique was developed for mining educational data. The proposed FSSML-EDM model involves three major processes. At the initial stage, the presented FSSML-EDM model incorporates the design of CSO-FS technique for electing feature subsets. Next, ELM classifier can be employed for the effective identification and classification of educational data. Finally, the AHB algorithm is utilized for effectively adjusting the parameters involved in the ELM model. The performance study revealed the effectual outcomes of the FSSML-EDM model over the compared models under several dimensions. Therefore, the FSSML-EDM model can be used as an effectual tool for EDM. In the future, feature reduction and outlier removal models can be employed to improve performance. In addition, the proposed model is presently tested on small-scale dataset, which needs to be explored. As a part of the future scope, the performance of the proposed model will be evaluated on a large-scale real-time dataset.

Author Contributions

Conceptualization, M.H.; Data curation, I.H.-J.; Formal analysis, I.H.-J.; Investigation, M.M.K.; Methodology, M.H. and R.F.M.; Project administration, M.M.K.; Resources, B.M.E.E.; Software, B.M.E.E.; Supervision, S.A.-K.; Validation, S.A.-K.; Visualization, S.A.-K.; Writing—original draft, M.H.; Writing—review & editing, R.F.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R125), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia; Also, the authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: (22UQU4400271DSR07).

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through contributions of all authors. All authors have given approval to the final version of the manuscript.

References

  1. Prakash, B.A.; Ramakrishnan, N. Leveraging Propagation for Data Mining: Models, Algorithms and Applications. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 2133–2134. [Google Scholar]
  2. Jalota, C.; Agrawal, R. Analysis of Educational Data Mining Using Classification. In Proceedings of the 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon), Faridabad, India, 14–16 February 2019; pp. 243–247. [Google Scholar] [CrossRef]
  3. Kenthapadi, K.; Mironov, I.; Thakurta, A.G. Privacy-preserving data mining in industry. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, WSDM’19: The Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, VIC, Australia, 11–15 February 2019; pp. 840–841. [Google Scholar]
  4. Yan, D.; Qin, S.; Bhattacharya, D.; Chen, J.; Zaki, M.J. 20th International Workshop on Data Mining in Bioinformatics (BIOKDD 2021). In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 4175–4176. [Google Scholar]
  5. Aldowah, H.; Al-Samarraie, H.; Fauzy, W.M. Educational data mining and learning analytics for 21st century higher education: A review and synthesis. Telemat. Inform. 2019, 37, 13–49. [Google Scholar] [CrossRef]
  6. Fernandes, E.; Holanda, M.; Victorino, M.; Borges, V.R.P.; Carvalho, R.; Van Erven, G. Educational data mining: Predictive analysis of academic performance of public school students in the capital of Brazil. J. Bus. Res. 2019, 94, 335–343. [Google Scholar] [CrossRef]
  7. De Andrade, T.L.; Rigo, S.J.; Barbosa, J.L.V. Active Methodology, Educational Data Mining and Learning Analytics: A Systematic Mapping Study. Inform. Educ. 2021, 20, 171–204. [Google Scholar] [CrossRef]
  8. Sáiz-Manzanares, M.; Rodríguez-Díez, J.; Díez-Pastor, J.; Rodríguez-Arribas, S.; Marticorena-Sánchez, R.; Ji, Y. Monitoring of Student Learning in Learning Management Systems: An Application of Educational Data Mining Techniques. Appl. Sci. 2021, 11, 2677. [Google Scholar] [CrossRef]
  9. Anand, N.; Sehgal, R.; Anand, S.; Kaushik, A. Feature selection on educational data using Boruta algorithm. Int. J. Comput. Intell. Stud. 2021, 10, 27. [Google Scholar] [CrossRef]
  10. Shrestha, S.; Pokharel, M. Educational data mining in moodle data. Int. J. Inform. Commun. Technol. (IJ-ICT) 2021, 10, 9. [Google Scholar] [CrossRef]
  11. Injadat, M.; Moubayed, A.; Nassif, A.B.; Shami, A. Systematic ensemble model selection approach for educational data mining. Knowl.-Based Syst. 2020, 200, 105992. [Google Scholar] [CrossRef]
  12. Ashraf, M.; Zaman, M.; Ahmed, M. An Intelligent Prediction System for Educational Data Mining Based on Ensemble and Filtering approaches. Procedia Comput. Sci. 2020, 167, 1471–1483. [Google Scholar] [CrossRef]
  13. Dabhade, P.; Agarwal, R.; Alameen, K.; Fathima, A.; Sridharan, R.; Gopakumar, G. Educational data mining for predicting students’ academic performance using machine learning algorithms. Mater. Today Proc. 2021, 47, 5260–5267. [Google Scholar] [CrossRef]
  14. Nahar, K.; Shova, B.I.; Ria, T.; Rashid, H.B.; Islam, A. Mining educational data to predict students performance. Educ. Inf. Technol. 2021, 26, 6051–6067. [Google Scholar] [CrossRef]
  15. Deb, S.; Gao, X.-Z.; Tammi, K.; Kalita, K.; Mahanta, P. Recent Studies on Chicken Swarm Optimization algorithm: A review (2014–2018). Artif. Intell. Rev. 2019, 53, 1737–1765. [Google Scholar] [CrossRef]
  16. Zhang, L.; Yang, H.; Jiang, Z. Imbalanced biomedical data classification using self-adaptive multilayer ELM combined with dynamic GAN. Biomed. Eng. Online 2018, 17, 181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Roozbeh, M.; Arashi, M.; Hamzah, N.A. Generalized Cross-Validation for Simultaneous Optimization of Tuning. Iran. J. Sci. Technol. Trans. A Sci. 2020, 44, 473–485. [Google Scholar] [CrossRef]
  18. Amini, M.; Roozbeh, M. Optimal partial ridge estimation in restricted semiparametric regression models. J. Multivar. Anal. 2015, 136, 26–40. [Google Scholar] [CrossRef]
  19. Roozbeh, M. Optimal QR-based estimation in partially linear regression models with correlated errors using GCV criterion. Comput. Stat. Data Anal. 2018, 117, 45–61. [Google Scholar] [CrossRef]
  20. Roozbeh, M.; Babaie-Kafaki, S.; Aminifard, Z. Improved high-dimensional regression models with matrix approximations applied to the comparative case studies with support vector machines. Optim. Methods Softw. 2022, 1–18. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Huang, C.; Huang, H.; Tang, S.; Dong, K. An optimization method: Hummingbirds optimization algorithm. J. Syst. Eng. Electron. 2018, 29, 386–404. [Google Scholar] [CrossRef]
  22. Duhayyim, M.A.; Marzouk, R.; Al-Wesabi, F.N.; Alrajhi, M.; Hamza, M.A.; Zamani, A.S. An Improved Evolutionary Algorithm for Data Mining and Knowledge Discovery. CMC-Comput. Mater. Contin. 2022, 71, 1233–1247. [Google Scholar]
Figure 1. Block diagram of FSSML-EDM technique.
Figure 1. Block diagram of FSSML-EDM technique.
Applsci 12 06787 g001
Figure 2. Framework of ELM.
Figure 2. Framework of ELM.
Applsci 12 06787 g002
Figure 3. Confusion matrix of FSSML-EDM technique on test data. (a) Entire dataset, (b) 70% of training dataset, and (c) 30% of testing dataset.
Figure 3. Confusion matrix of FSSML-EDM technique on test data. (a) Entire dataset, (b) 70% of training dataset, and (c) 30% of testing dataset.
Applsci 12 06787 g003
Figure 4. Result analysis of FSSML-EDM technique on entire dataset.
Figure 4. Result analysis of FSSML-EDM technique on entire dataset.
Applsci 12 06787 g004
Figure 5. Result analysis of FSSML-EDM technique on 70% of training dataset.
Figure 5. Result analysis of FSSML-EDM technique on 70% of training dataset.
Applsci 12 06787 g005
Figure 6. Result analysis of FSSML-EDM technique on 30% of testing dataset.
Figure 6. Result analysis of FSSML-EDM technique on 30% of testing dataset.
Applsci 12 06787 g006
Figure 7. Accuracy graph analysis of FSSML-EDM technique.
Figure 7. Accuracy graph analysis of FSSML-EDM technique.
Applsci 12 06787 g007
Figure 8. Loss graph analysis of FSSML-EDM technique.
Figure 8. Loss graph analysis of FSSML-EDM technique.
Applsci 12 06787 g008
Figure 9. Precision-recall curve analysis of FSSML-EDM technique.
Figure 9. Precision-recall curve analysis of FSSML-EDM technique.
Applsci 12 06787 g009
Figure 10. ROC curve analysis of FSSML-EDM technique.
Figure 10. ROC curve analysis of FSSML-EDM technique.
Applsci 12 06787 g010
Figure 11. P r e c n , r e c a l , and a c c y analysis of FSSML-EDM technique.
Figure 11. P r e c n , r e c a l , and a c c y analysis of FSSML-EDM technique.
Applsci 12 06787 g011
Figure 12. F s c o r e , M C C , and k a p p a analysis of FSSML-EDM technique.
Figure 12. F s c o r e , M C C , and k a p p a analysis of FSSML-EDM technique.
Applsci 12 06787 g012
Table 1. Dataset details.
Table 1. Dataset details.
DetailsValues
Number of samples649
Feature count32
Class count2
Sourcehttps://archive.ics.uci.edu/ml/datasets/student+performance (accessed on 30 January 2022)
Table 2. Result analysis of FSSML-EDM technique with distinct measures and datasets.
Table 2. Result analysis of FSSML-EDM technique with distinct measures and datasets.
Class LabelsAccuracyPrecisionRecallF-ScoreMCCKappa Score
Entire Dataset
Pass97.2397.5099.2798.3889.08-
Fail97.2395.5686.0090.5389.08-
Average97.2396.5392.6494.4589.0888.91
Training Set (70%)
Pass97.5897.5099.7498.6189.57-
Fail97.5898.1584.1390.6089.57-
Average97.5897.8291.9494.6089.5789.22
Testing Set (30%)
Pass96.4197.4898.197.7988.22-
Fail96.4191.6789.1990.4188.22-
Average96.4194.5893.6594.188.2288.2
Table 3. Comparative analysis of FSSML-EDM approach with existing methods [21].
Table 3. Comparative analysis of FSSML-EDM approach with existing methods [21].
MethodsPrecisionRecallAccuracyF-ScoreMCCKappa
FSSML-EDM 94.5893.6596.4194.1088.2288.20
IEAFSSNFC93.8192.3990.3393.0173.7873.37
NFC Algorithm81.7691.4181.6686.9956.2150.10
NN Algorithm67.5885.7664.3076.0357.5267.82
SVM Algorithm68.8386.7766.3676.6366.2766.17
DT Algorithm78.2293.4077.8183.0252.8759.80
RF Algorithm74.9591.9075.4083.3962.6858.34
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hamdi, M.; Hilali-Jaghdam, I.; Khayyat, M.M.; Elnaim, B.M.E.; Abdel-Khalek, S.; Mansour, R.F. Chicken Swarm-Based Feature Subset Selection with Optimal Machine Learning Enabled Data Mining Approach. Appl. Sci. 2022, 12, 6787. https://doi.org/10.3390/app12136787

AMA Style

Hamdi M, Hilali-Jaghdam I, Khayyat MM, Elnaim BME, Abdel-Khalek S, Mansour RF. Chicken Swarm-Based Feature Subset Selection with Optimal Machine Learning Enabled Data Mining Approach. Applied Sciences. 2022; 12(13):6787. https://doi.org/10.3390/app12136787

Chicago/Turabian Style

Hamdi, Monia, Inès Hilali-Jaghdam, Manal M. Khayyat, Bushra M. E. Elnaim, Sayed Abdel-Khalek, and Romany F. Mansour. 2022. "Chicken Swarm-Based Feature Subset Selection with Optimal Machine Learning Enabled Data Mining Approach" Applied Sciences 12, no. 13: 6787. https://doi.org/10.3390/app12136787

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop