You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

11 October 2022

A Method for Predicting the Academic Performances of College Students Based on Education System Data

,
and
School of Information Engineering, Shenyang University, Shenyang 110044, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Applications of Data Mining Algorithms and Big Data Analytics in Education

Abstract

With the development of university campus informatization, effective information mined from fragmented data can greatly improve the management levels of universities and the quality of student training. Academic performances are important in campus life and learning and are important indicators reflecting school administration, teaching level, and learning abilities. As the number of college students increases each year, the quality of teaching in colleges and universities is receiving widespread attention. Academic performances measure the learning ‘effects’ of college students and evaluate the educational levels of colleges and universities. Existing studies related to academic performance prediction often only use a single data source, and their prediction accuracies are often not ideal. In this research, the academic performances of students will be predicted using a feedforward spike neural network trained on data collected from an educational administration system and an online learning platform. Finally, the performance of the proposed prediction model was validated by predicting student achievements on a real dataset (involving a university in Shenyang). The experimental results show that the proposed model can effectively improve the prediction accuracies of student achievements, and its prediction accuracy could reach 70.8%. Using artificial intelligence technology to deeply analyze the behavioral patterns of students and clarify the deep-level impact mechanisms of the academic performances of students can help college educators manage students in a timely and targeted manner, and formulate effective learning supervision plans.

1. Introduction

Due to the widespread use of educational information management systems in colleges and universities, the amount of data in databases has grown, and the data held by colleges and universities have become more abundant [1]. Unfortunately, the use of this part of the data has not been further studied, it is still only for the purpose of providing simple inquiries and statistical reports. It is impossible to obtain implicit value information from a large amount of data, such as the relationship between students’ grades and daily behavior, where students go after graduation, etc. [2]. How to effectively use these data could effectively guide college administrators, improve the teaching management levels at colleges and universities, and further improve the quality of running schools; these will be key issues that college and university administrators will consider in the future.
In college and university educational affairs activities, student achievements are important for evaluating knowledge; they are also key factors for evaluating teaching quality [3]. With the current open teaching management model in colleges and universities, and the rapid development of online education, coupled with the outbreak of COVID-19 (coronavirus disease 2019), there are more data on online learning behaviors. This allows students to have a higher degree of learning freedom; students are free to set their plans and goals, which also leads to more factors that affect academic performance [4]. The changing laws of academic performances are discovered from the information and associations behind data, such as in-depth research, factor analyses that affect student performances, and artificial intelligence algorithms used to find the key factors affecting academic performances [5]. Thus, it is practical to analyze students’ grades and predict future grades. Changes in grades can provide suggestions for teaching, improve education and management, and improve learning efficiencies. Therefore, this is of great significance in theory and practical applications.
Academic achievement prediction refers to predicting a student’s future academic performance based on existing behavioral data. Academic achievement prediction is one of the most primitive and popular applications in the field of educational data mining [6]. Specifically, academic performance prediction generally predicts the development trend of academic performance by collecting the behavioral data of students and analyzing behavioral patterns using statistical or data mining methods. The role of academic performance prediction is to predict the development trend of academic performance in advance and find out the possible problems of students in time. This allows students and teachers to react in a timely manner and prevent the risk of failing the course. As the educational environment continues to change, so does research in this field [5]. Academic achievement prediction is an important research problem in the field of educational data mining, which has been intensively studied by many researchers [7]. However, many challenges remain in terms of prediction accuracy and interpretability due to the lack of richness and diversity of data sources and related features. Some representative works are summarized as follows. Qu et al. [8] proposed a student achievement predicting framework, which includes data processing and student achievement predicting. The proposed method has a better performance in comparison to support vector machines, naïve Bayes, logistic regression, and multi-layer perception. Kostopoulos [9] presented a novel semi-supervised regression algorithm for predicting the final grade of undergraduate students in a distance online course. Yang et al. [10] predicted course achievements in Moodle through the procrastination behaviours of students using their homework submission data. A Moodle course from the University of Tartu in Estonia was used, which included 242 students to assess the efficacy of the proposed approach. Tomasevic et al. [7] provided a comprehensive analysis and comparison of the state-of-the-art supervised machine learning techniques used to solve exam performance predictions. Nieto-Reyes et al. [11] proposed a methodology that automates the prediction of academic performances at the end of the course using data recorded during the first tasks of the academic year. This methodology allows teachers and evaluators to define the variables that they consider most appropriate to measure those aspects related to the academic performances of students. Miao et al. [12] improved the neural network algorithm, and combined support vector machines to build a student grade-prediction model; they used the principle component analysis to reduce the dimensionality of the sample data. Ali et al. [13] predicted the achievement of 82 undergraduates enrolled in a hybrid English for Business Communication course using clustering techniques. Classification results revealed that log data and engagement activities successfully predicted academic performances with more than 88% accuracy. Baashar et al. [14] examined and surveyed the current literature regarding the ANN methods used in predicting academic performances; they attempted to capture a pattern of the most used artificial neural network techniques and algorithms. Li et al. [15] proposed an end-to-end deep learning model that automatically extracts features from students’ multi-source heterogeneous behavior data to predict academic performances. Liu et al. [16] proposed a student achievement prediction model based on the evolutionary spiking neural network. The experimental results show that the proposed method can effectively improve the prediction accuracies of student achievements.
Although the above models have been successfully applied to student achievement predictions and these experiences are very helpful for improving the accuracy of student achievement predictions, these prediction models still have some room for improvement in the design stage and implementation of actual teaching data prediction, such as limited data sources, prediction low precision, poor efficiency, etc. Therefore, this paper mainly explores a student achievement prediction model based on the feedforward spiking neural network. The learning prediction model was deeply studied to improve the generalization performance and prediction accuracy of the feedforward spiking neural network, and a real dataset was analyzed from student information and course information data (from the 19th-grade IoT at a university in Shenyang, China). Finally, the proposed model was applied to student grade prediction on the dataset and compared with state-of-the-art algorithms to verify its effectiveness. The experimental results show that the proposed algorithm is better than similar learning algorithms in the accuracy of student achievement prediction.
The main contributions of this paper are as follows.
  • This paper processed real student and course data from a university in Shenyang and used it for the training and testing of the predicted model.
  • A feedforward spiking neural network model is proposed to predict students’ grades; the simulation results show the advantages of the proposed model using student grade data and course information data in predicting real students’ grades.
  • It is helpful for teachers to implement timely intervention and for students to adjust their learning statuses, which is of great significance to the harmonious development of teaching and learning.
The remainder of this paper is summarized below. Section 2 describes the concept of student achievement prediction, pattern classification, and spiking neural network. Section 3 describes the content related to the real dataset from a university in Shenyang. Section 4 describes how to design the proposed model in detail, including encoding and algorithm designs. Section 5 presents the results of the experimental evaluation. The proposed model is compared with the state-of-the-art models, and the simulation results are evaluated and discussed. Finally, Section 6 presents the conclusions and outlines future research lines.

3. Real Datasets in Education Systems

In this section, we will use the dataset in the educational administration system of all students in the 19th-grade of the Internet of Things major from a university in Shenyang, and course learning data from the Chaoxing Xuexitong system to analyze and study the factors that have the greatest impacts on the final grades of students. Based on this dataset, a suitable feedforward spiking neural network prediction model is established to predict student achievements.
With the popularization of information technology, more data in the school’s education information system can be used for research, and the data can be deeply excavated to discover the inherent laws of the data itself, to help decision-makers better manage schools [31]. Early research on student achievement prediction focused on education and psychology, using data mostly from questionnaires or student self-reports. Different from previous studies, we obtained datasets for predicting student performances through the school’s educational administration system, the online learning platform of the Chaoxing Xuexitong system, and the enrollment information statistics table. From a macro perspective, the mining of these data is very valuable. If the data can be correlated, many research questions can be generated. After removing the private data of the student’s identity, we can match the data with the student’s identity one-to-one, to better manage and supervise the students’ studies effectively.
Throughout the student achievement prediction modeling, the first thing to do is preprocess the data from different sources in the information systems [32]. Data preprocessing takes about 40% of the time, and the quality of data processing directly affects the accuracy of the model. The preprocessed data not only save a lot of space and time, but they also obtain the prediction results that have a better role in decision-making and prediction. Secondly, due to the different sources of educational data, the dimensions of attributes and features are inconsistent, and there is a lack of data features. In order to obtain better-quality modeling data, certain data integration, cleaning, regularization, and transformation must be performed [33]. More specifically, data integration is a technique that involves combining data from multiple data sources (databases, data cubes, or generic files) into a consistent data store (such as a data warehouse). Since the data integration of different disciplines involves different theoretical foundations and rules, data integration can be said to be a difficult point in data preprocessing. Naming rules and requirements for each data source may be inconsistent. To extract data from multiple data sources into a data warehouse, in order to ensure the accuracy of the experimental results, all data formats must be unified. There are two broad categories of methods for implementing a unified format. One is to modify each data source first, and then extract it uniformly into the data warehouse. Data cleaning is the most time-consuming and tedious, but also the most important step in the data preparation process. This step can effectively reduce the problem of conflicting situations that may arise during the learning process. Data transformation is the use of linear or nonlinear mathematical transformation methods to compress multi-dimensional data into data with fewer dimensions and eliminate their differences in characteristics, such as space, attributes, time, and precision. Although such methods are usually lossy to the original data, the results tend to have greater utility.

4. Proposed Method

If the examination is one of the most important means to evaluate the quality of college training, then the course’s score is the most direct factor to evaluate the quality of students. If students can predict their grades in advance based on their previous academic performances, and this reminds them to study hard to avoid failing, this is of great significance in education and teaching research [34]. Figure 2 shows the overall experimental process scheme for the construction of the experimental model of student achievement predictions based on the spiking neural network. The proposed feedforward spiking neural network mainly consists of three parts: encoding the input data, training and learning the spiking neural network, and decoding and outputting the learning results, as shown in Figure 2. The model mainly realizes the classifications of students’ grades, i.e., predicts the course grades of students in three grades.
Figure 2. Structure topology of spiking neural network based on student achievement prediction.
According to Figure 2, the implementation process of the proposed feedforward spiking neural network is shown in Figure 3. The first spiking coding is used to encode the student data and course data into the input spiking sequence and the expected spiking sequence; the spiking sequence is passed to the hidden layer, where the learning rate and the synaptic time delay are adaptively adjusted and the information is transmitted to the output layer. The output layer outputs the actual spiking sequence and compares it with the expected output spiking sequence to generate an error rate, and then back-propagates to adjust the learning according to the value of the error rate and synaptic time delay, thereby reducing the error value until the error is less than a certain limit or the iteration number reaches the required number of iterations to stop learning, and decode the output spike sequence into classification categories.
Figure 3. The architecture of student achievement prediction model.
To simplify the operation of spiking neural networks, the weight-learning rules in this paper were learned based on gradient descent, which can speed up network training while the spiking neural network is running. Gradient descent learning rules are the most commonly used and relatively classic and mature algorithms in neural network learning. The idea is to construct the loss function of the spiking neural network, and use the principle of gradient descent to obtain the optimal value of the loss function, and the rules of the gradient descent learning algorithm of the spiking neural network are similar to the principle of the artificial neural network. The difference between the target output and the actual output is used to construct a loss function, and then the error is backpropagated to change the weights in the network. The adjustment of network weights can be expressed as Equation (1).
Δ w = η E = η E w
where η represents the learning rate of the spiking neural network. E represents the partial derivative of the loss function E to the weight w in the spiking neural network, and the gradient value of the loss function is adjusted to the weight.
The specific expression of the error loss function E is shown in Equation (2).
E = 1 2 f = 1 N ( t f t f ) 2
where t f represents the actual excitation time of the f-th spike in the spiking neural network; t f is the expected output time of the f-th spike in the spiking neural network; N is the number of spikes fired by a neuron in a spiking neural network.
Gradient descent learning algorithms should budget based on the neuron’s internal state variables. This paper uses spike response mode (SRM) 0, which is a simplified model of the SRM model [35], and its specific internal state variables are shown in Equation (3).
u j ( t ) = i k w i j k ε ( t t i d i j k ) = i k w i j k y i j k
where i is the number of spikes in the pre-synaptic neuron and y i j k is the symbolic representation of the activation function.
The training rule is shown in Equation (4).
d i = α 2 f = 1 N ( ( t f t f ) 2 ) d i = α 2 f = 1 N ( ( t f t f ) 2 ) t f t f d i = α f = 1 N ( t f t f ) t f d i
where d i represents the adjustment amount of the i-th synaptic time delay of a neuron in the spiking neural network. α is similar to the learning rate in weight adjustment, indicating the magnitude of each time delay adjustment. Only the t f d i item needs to be calculated separately, and the rest are obtained in the intermediate calculation process of the weights.
The calculation method of t f d i is also similar to the weight adjustment. It adopts the recursive theory based on the gradient descent method. The adjustment result of the synaptic time delay in the next iteration has an important relationship with the spike time of the previous iteration process. The formula is shown in Equation (4).
t f d i = t f u ( t f ) ( u ( t f ) d i + u ( t f ) t ( f 1 ) t ( f 1 ) d i ) t ( f 1 ) d i = t ( f 1 ) u ( t ( f 1 ) ) ( u ( t ( f 1 ) ) d i + u ( t ( f 1 ) ) t ( f 2 ) t ( f 2 ) d i ) t 1 d i = t 1 u ( t 1 ) u ( t 1 ) d i
In Formula (4), only the u ( t f ) d i item needs to be calculated separately; the rest of the calculation results are involved in the weight calculation process, and they need to be called directly. This operation both facilitates computations and adjusts synaptic time delays with minimal effort. u ( t f ) d i can be calculated according to Equation (5).
u ( t f ) d i = ( i = 1 N w i ε ( t d i ) ) d i = w i ( i = 1 N w i ε ( t f t i g d i ) ) d i = w i i = 1 N t f t i g d i τ 2 e 1 t f t i g d i τ e 1 t f t i g d i τ τ = w i i = 1 N τ ( t f t i g d i ) ( 1 τ 1 t f t i g d i )
In order to ensure the accuracy of the training and testing of the proposed feedforward spiking neural network and its authenticity, this paper divides the student data set into two parts, one for training and one for testing. In order to better explain the proposed model, the pseudo-code of its specific algorithm is shown in Algorithm 1.
Algorithm 1 The pseudo-code of the proposed algorithm for student achievement prediction.
Require: Student dataset and course dataset.
Ensure: Predicted results.
  1:
Load the student dataset and course dataset, and initialize the weights and synaptic delays of the spiking neural network.
  2:
Set the parameters of the spiking neural network, including the initial learning rate L 0 , time step t, different time constants τ , threshold ϑ , the number of neurons in the output layer o _ n , the number of neurons in the input layer and hidden layer h _ n , and the connections between neurons number c _ n and the number of synaptic time delays d in each connection.
  3:
Randomly generate initial weights and synaptic time delays, and normalize these parameters.
  4:
Start training the spiking neural network and forward the spiking information to the next neuron. Computation of the internal state variables requires a comparison of the postsynaptic membrane potential with the thresholds of the spiking neural network. If the value of the internal state variable is greater than the threshold, a spike signal is generated and the time of the spiking emission is recorded.
  5:
Use the actual spike time and the expected spike time to construct an error function, and judge the updated range of the learning rate according to the error value.
  6:
According to the obtained new learning rate L, back-propagation is performed, and the gradient descent principle is used to change the weight of the output layer and the hidden layer of the spiking neural network, so that the error gradually becomes smaller.
  7:
Repeat the above process until the error is less than the minimum error of the original algorithm or meet a certain number of iterations.
  8:
Decode the network output.

5. Experimental Studies

In order to verify the effect of the student achievement prediction model proposed in this paper, an experiment was designed to compare the proposed model with other similar models, and the final exam results of students were predicted. The prediction task is to predict the grades based on both the student and some of the courses, and the prediction results are accurate concerning whether the student can pass a course or the possible test levels of a certain course. The data and comparison methods used in the experiments are described below. First, we describe the real education datasets and the evaluation indicator used in the experiment. All experiments are based on the real education dataset of a university in Shenyang city, Liaoning Province, China. Second, we give a brief comparison of the five experimental models and provide the experimental setup used in this study. These models include the decision tree, random forest, neural network, XGBoost, and SVM. Finally, the comparison results of these experimental models under the real education datasets are discussed.

5.1. Experimental Datasets and Experimental Conditions

5.1.1. Datasets

In order to verify the validity of the proposed model, we collected and sorted out two datasets, corresponding to the student datasets and course datasets of the 19th-grade Internet of Things major at a university in Shenyang city, Liaoning Province, China. The features used in the experiments include basic attribute information of students and courses. We deciphered the student data and course data; there were a total of 3976 pieces of processed data. Table 1 gives the basic information about the above two datasets, respectively. There are 55 students in the dataset, and it covers student performance data for 62 courses in 6 semesters. The dataset contains 26 variables related to student grades. The size of the dataset is 3976 entries, which includes the results from normal exams (3410) and make-up exams (566).
Table 1. Student Personal Information.
The dataset contains 55 student records from different genders for 6 semesters, consisting of 26 information properties, including student ID, grade point average (GPA), entrance grades, semester, course attributes, performance in class, etc. About 70 percent of students came from the province where the university was located. This paper analyzed these datasets and we built a model to predict student grades. Through the modeling results, it was discovered that there were imbalanced data in the dataset. The imbalance of data makes the sample size of some rare attribute values obtained in the training set more sparse, which further exacerbates the sparsity of sample features. To address the problem of data imbalance, we divide the predicted grades into three levels [‘L’, ‘M’, ‘H’], which will be used as the standard for judging students’ grades. Among them, ‘L’ (0–64) means a failing score, ‘M’ (65–84) means a medium score, and ‘H’ (85–100) means a high score. The datasets were modeled under three-level classification tasks in Table 2. The original value in Table 2 is the real value in the dataset. The classification in Table 2 refers to the three values of the three-level classification system on a 100-point scale. The encoding in Table 2 represents the three-category numerical value for the convenience of the programming.
Table 2. The three-level classification system.
Figure 4 shows the proportion of students by gender in the real educational datasets, of which 32 are boys and 23 are girls.
Figure 4. Statistics of Student Gender.
The student enrollment data table includes information, such as college entrance examination scores, class, major, grade, no. (student ID), gender, place of origin, political status, graduate high school, date of birth, etc. It was found through the experiments that college entrance examination scores, gender, political affiliation, grades, and other information greatly influence the results of the scores. The experimental data in Figure 5 shows that most student scores are distributed between 500 points, some of the scores were between 500 and 520, and the number of students below 500 points is very small.
Figure 5. Student performance analysis entrance examination.
As can be seen from Figure 6, most students can obtain good grades in most courses. Among them, the number of courses with good grades was 52.82% of the total number of courses, the number of courses with middle grades was 39.89% of the total number of courses, and the courses with poor grades accounted for 7.29% of the total number of courses. This result shows that most students could master the knowledge points of the course and achieve ideal grades.
Figure 6. Proportional distribution of course grades.
In order to further analyze the behaviors of these students, a questionnaire for statistical work and postgraduate entrance examination was set up. The results of the questionnaire are shown in Figure 7. Observing Figure 7, it is not difficult to find that the number of boys who chose to work; graduate school was the same, while the number of girls who chose to work was more.
Figure 7. The relationship between gender and the proportion of graduate career choices.
Figure 8 shows the proportion of students’ political affiliations. It can be seen from Figure 8 that most of the students were members of the Communist Youth League, and the number of male party members was higher than the number of female party members.
Figure 8. The relationship between gender and the proportion of party members.
The overall ranking of the 55 students is shown in Figure 9. The comparative relationship between the overall grades of college courses and the college entrance examination scores of 55 students. The purple dotted line in the middle of Figure 9 represents the results of the college entrance examination. It can be seen from the figure that the scores of the students in the college entrance examination are relatively similar. However, the blue star line indicates the ranking of all university courses. From the results in Figure 9, it can be seen that after entering the university, the grades among students gradually opened up. This has a lot to do with the degree of effort and the major interest of the students after admission.
Figure 9. Overall ranking of 55 students.

5.1.2. Experimental Conditions

In order to verify the advantages of the proposed model in predicting student achievement, this paper selects five machine learning models for comparison. These models include the decision tree, random forest, neural network, XGBoost, and SVM. They were employed to evaluate the performance of the proposed model.
These experimental models were carried out on a 64-bit platform with a Windows 10 operating system, which consists of memory 12GB, the processor with Intel i7-4710MQ (the main frequency of 2.50 GHz), and the software used in the PyCharm community with Python 3.10. In addition, due to the different dimensions in the dataset, there is a certain degree of a sample annihilation problem. Before executing these experimental models, this paper adopted the normalization technique to process the datasets.
Since the student achievement data obtained in this paper contained a certain amount of noise, the generalization error could not be directly used as a signal to understand the generalization ability of the prediction model. It is very expensive to reciprocate between the deployment environment and the training model, and the model cannot be used for the training dataset. The degree of fit is used as a signal to understand the generalization ability of the model. Therefore, a better way is to split the student achievement dataset into two parts, the training dataset, and the test dataset. We can use the training dataset to train the model, and then use the error on the test dataset as the generalization error of the final model in response to real-world scenarios. With the test dataset, we want to verify the final effect of the model. We only need to calculate the error of the trained model on the test dataset, and we can consider this error to be an approximation of the generalization error. We only need to let our trained model on the test dataset be the smallest error. The distributions of the training dataset and test dataset of the student achievement dataset in this paper are shown in Table 3.
Table 3. The distribution of the training dataset and test dataset of the student achievement dataset.

5.2. Comparing the Results of All Experimental Algorithms

The normalized dataset was selected to evaluate the predictive ability of the proposed model for student achievement and compare it with these experimental models. Note that we converted student grades to ‘H’, ‘M’, and ‘L’ according to the rules in Table 2.

5.2.1. Comparing Results with All Experimental Models on the Real Educational Datasets

Table 4 shows the prediction results of all experimental models on the real educational dataset. Observing the experimental results in Table 4, it was concluded that the results of the proposed model were significantly better than those of other models. From the experimental results in Table 4, it can be seen that random forest had the worst results. The proposed model outperformed the comparison models in the experiments in four metrics precision, including recall, f1-score, and accuracy. It is particularly worth noting that the proposed model achieved good results in that the prediction ability on ‘L’ was significantly better than other models. This led to an overall improvement in the performance of the proposed model. This result is related to the fact that ‘L’ belongs to small samples, which illustrates the predictive ability of the proposed model on small sample data.
Table 4. Comparing the results with all experimental algorithms on the ‘xAPI-Edu-data’ dataset.
To further analyze the predictive ability of the proposed algorithm, Figure 10 presents the prediction results of each model. Figure 10 shows the performance prediction results of each model on the test dataset. It is easy to find that most models show poor results in the 100-point prediction. The experimental results show that this is related to factors, such as data imbalance and incomplete student data attributes.
Figure 10. Predicted results of percentile grades.

5.2.2. Discussions

From the results in Table 4, we can see that the proposed model achieves good results in most of the experimental results. The feedforward spiking neural network model designed in this paper further improves the prediction performance, indicating that the model is effective and can better predict students’ final grades. Based on the above experimental results, we can conclude that the proposed model is effective to predict student achievement.
Through the development of the above experiments and the implementation of the practical research on the prediction of academic performance based on the feedforward spiking neural network, some theoretical and practical research results have been achieved. However, due to the limitations of some factors, this study has the following three shortcomings.
  • We carried out a study on the prediction of academic performances on the basis of the feedforward spike neural network but did not try more new intelligent analysis techniques; the selected case samples are expected to be further expanded and judged through more evaluation index comparisons, the accuracy, and validity of the learning performance prediction model, and the practical value of teaching.
  • The deep relationship between portrait application and academic performance prediction is expected to be further explored. Since this study focused on building a learning performance prediction model, the dimensions of the learner portrait construction and the corresponding data indicators were very rich (which affect the accuracy of learning achievement prediction). On the one hand, it still needed to be further verified and revised to improve the support efficiency of learner portraits for teaching and learning; on the other hand, the prediction effect of academic performance can be improved by optimizing the spiking neural network model, and the key factors of academic performance prediction can be further explored to help teachers and students. It can provide more accurate and personalized learning services.
  • Experimental data samples were limited since the learner samples selected in this study were from two classes in one grade; for six-semester courses, the acquisition of learning data was also limited by the online learning platform and educational administration system. These data were still shallow and single, and the data amounts were limited. They were also relatively sparse, so the prediction accuracy of student grades still needed to be further improved. In the follow-up, on the one hand, we will combine the latest research results of behavioral science and brain science to collect more data on the learning processes of learners.

6. Conclusions

In a digital campus environment, the multi-source heterogeneous data of college students is integrated to break data barriers between different online and offline learning spaces and provide theoretical and technical bases for the research on the trend prediction of college students’ academic performances [36]. It is very important to mine effective information from scattered information data to promote the development of campus activities. Student performance is the most important factor to measure a student’s personal ability and a teacher’s teaching level [37]. Predicting a student’s course performance in advance can provide early warnings and timely correction roles for students and teachers. Student performance prediction is also the most important direction in educational data mining.
This paper mainly studied the relationship between the content of the courses that students study and the test scores. Based on the analysis results, knowledge features are used to predict academic performances, and a new model for predicting student performances based on the feedforward spiking neural network was designed. This paper studied the prediction of follow-up grades through the relevant data information and course information of students. In order to verify the accuracy of the performance prediction model proposed in this paper, this paper mainly completed the following work. First of all, regarding data preprocessing, due to the selection of real college student achievement data, it is necessary to perform operations, such as data normalization and default value processing. Secondly, the influence of a single feature on academic performance was analyzed, and the features that had greater impacts on the academic performances were screened. Third, the academic performance was predicted based on the feedforward spiking neural network prediction model. At the same time, the optimal academic achievement model was constructed through parameter tuning. Finally, the advantages of the proposed feedforward spiking neural network in predicting academic performances were demonstrated through simulation experiments. The research data came from the personal statistics and student transcripts of 19th-grade IoT students at the School of Information Engineering (at a university in Shenyang). Based on these data, this paper proposes a model to predict student performances. The experiments in this paper were based on real educational datasets. Through a large number of experiments and result analyses, the effectiveness of the curriculum knowledge characteristics and prediction model obtained by the feedforward spiking neural network proposed in this paper were verified in predicting the learning performances.
This paper systematically summarizes the predictors of academic performance, which could help improve the efficiency and quality of academic performance predictions. However, the inductive perspectives of the academic achievement predictors used in this paper are relatively simple, and there is a lack of standardized data selection methods. The perspective of academic achievement prediction is course-centered and student-centered and lacks the behavioral characteristics of some students. Furthermore, the research focuses on evaluating the performances of forecasting algorithms, and it is easy to overlook the importance of rationally choosing data metrics. The relationship between student portraits and academic performance prediction needs to be further explored. Student portraits can describe the characteristics of students at different learning stages, and learning outcomes can standardize and quantify learning outcomes. Combined with keywords such as “learning/academic performance prediction”, there are few detailed studies in this field, indicating that the relationship between student portraits and academic performance prediction research is not close, and the advantages of student portraits in future research will depend on the performance improvements in prediction [38].

Author Contributions

Conceptualization, C.L. and Z.Y.; methodology, C.L.; software, H.W.; validation, C.L., H.W. and Z.Y.; writing—original draft preparation, C.L.; writing—review and editing, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by 69 batches of general funding projects from the China Postdoctoral Science Foundation, China (grant no. 2021M693858), and Technological Innovation Program for Young People of Shenyang City, China (grant no. RC210400), and Scientific Research Funding Project of the Education Department of Liaoning Province, China (grant no. 2020JYT05).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editors and reviewers for providing useful comments and suggestions to improve the quality of this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Okewu, E.; Adewole, P.; Misra, S.; Maskeliunas, R.; Damasevicius, R. Artificial Neural Networks for Educational Data Mining in Higher Education: A Systematic Literature Review. Appl. Artif. Intell. 2021, 35, 983–1021. [Google Scholar] [CrossRef]
  2. Asif, R.; Merceron, A.; Ali, S.A.; Haider, N.G. Analyzing undergraduate students’ performance using educational data mining. Comput. Educ. 2017, 113, 177–194. [Google Scholar] [CrossRef]
  3. You, J.W. Identifying significant indicators using LMS data to predict course achievement in online learning. Internet High. Educ. 2016, 29, 23–30. [Google Scholar] [CrossRef]
  4. González-Calatayud, V.; Prendes-Espinosa, P.; Roig-Vila, R. Artificial Intelligence for Student Assessment: A Systematic Review. Appl. Sci. 2021, 11, 5467. [Google Scholar] [CrossRef]
  5. Yang, F.; Li, F.W.B. Study on student performance estimation, student progress analysis, and student potential prediction based on data mining. Comput. Educ. 2018, 123, 97–108. [Google Scholar] [CrossRef]
  6. Dutt, A.; Ismail, M.A.; Herawan, T. A Systematic Review on Educational Data Mining. IEEE Access 2017, 5, 15991–16005. [Google Scholar] [CrossRef]
  7. Tomasevic, N.; Gvozdenovic, N.; Vranes, S. An overview and comparison of supervised data mining techniques for student exam performance prediction. Comput. Educ. 2020, 143, 103676. [Google Scholar] [CrossRef]
  8. Qu, S.; Li, K.; Zhang, S.; Wang, Y. Predicting Achievement of Students in Smart Campus. IEEE Access 2018, 6, 60264–60273. [Google Scholar] [CrossRef]
  9. Kostopoulos, G.; Kotsiantis, S.; Fazakis, N.; Koutsonikos, G.; Pierrakeas, C. A Semi-Supervised Regression Algorithm for Grade Prediction of Students in Distance Learning Courses. Int. J. Artif. Intell. Tools 2019, 28, 1940001. [Google Scholar] [CrossRef]
  10. Yang, Y.; Hooshyar, D.; Pedaste, M.; Wang, M.; Huang, Y.M.; Lim, H. Predicting course achievement of university students based on their procrastination behaviour on Moodle. Soft Comput. 2020, 24, 18777–18793. [Google Scholar] [CrossRef]
  11. Nieto-Reyes, A.; Duque, R.; Francisci, G. A Method to Automate the Prediction of Student Academic Performance from Early Stages of the Course. Mathematics 2021, 9, 2677. [Google Scholar] [CrossRef]
  12. Miao, J. A hybrid model for student grade prediction using support vector machine and neural network. J. Intell. Fuzzy Syst. 2021, 40, 2673–2683. [Google Scholar] [CrossRef]
  13. Ali, A.D.; Hanna, W.K. Predicting Students’ Achievement in a Hybrid Environment Through Self-Regulated Learning, Log Data, and Course Engagement: A Data Mining Approach. J. Educ. Comput. Res. 2022, 60, 960–985. [Google Scholar] [CrossRef]
  14. Baashar, Y.; Alkawsi, G.; Mustafa, A.; Alkahtani, A.A.; Alsariera, Y.A.; Ali, A.Q.; Hashim, W.; Tiong, S.K. Toward Predicting Student’s Academic Performance Using Artificial Neural Networks (ANNs). Appl. Sci. 2022, 12, 1289. [Google Scholar] [CrossRef]
  15. Li, X.; Zhang, Y.; Cheng, H.; Li, M.; Yin, B. Student achievement prediction using deep neural network from multi-source campus data. Complex Intell. Syst. 2022, 1–14. [Google Scholar] [CrossRef]
  16. Liu, C.; Wang, H.; Du, Y.; Yuan, Z. A Predictive Model for Student Achievement Using Spiking Neural Networks Based on Educational Data. Appl. Sci. 2022, 12, 3841. [Google Scholar] [CrossRef]
  17. Sekeroglu, B.; Abiyev, R.; Ilhan, A.; Arslan, M.; Idoko, J.B. Systematic Literature Review on Machine Learning and Student Performance Prediction: Critical Gaps and Possible Remedies. Appl. Sci. 2021, 11, 10907. [Google Scholar] [CrossRef]
  18. Cogliano, M.; Bernacki, M.L.; Hilpert, J.C.; Strong, C.L. A Self-Regulated Learning Analytics Prediction-and-Intervention Design: Detecting and Supporting Struggling Biology Students. J. Educ. Psychol. 2022. [Google Scholar] [CrossRef]
  19. Tsai, Y.h.; Lin, C.h.; Hong, J.c.; Tai, K.h. The effects of metacognition on online learning interest and continuance to learn with MOOCs. Comput. Educ. 2018, 121, 18–29. [Google Scholar] [CrossRef]
  20. Steinmayr, R.; Weidinger, A.F.; Wigfield, A. Does students’ grit predict their school achievement above and beyond their personality, motivation, and engagement? Contemp. Educ. Psychol. 2018, 53, 106–122. [Google Scholar] [CrossRef]
  21. Fahd, K.; Venkatraman, S.; Miah, S.J.; Ahmed, K. Application of machine learning in higher education to assess student academic performance, at-risk, and attrition: A meta-analysis of literature. Educ. Inf. Technol. 2022, 27, 3743–3775. [Google Scholar] [CrossRef]
  22. Ahmad, S.; El-Affendi, M.A.; Anwar, M.S.; Iqbal, R. Potential Future Directions in Optimization of Students’ Performance Prediction System. Comput. Intell. Neurosci. 2022, 2022. [Google Scholar] [CrossRef]
  23. Alhassan, A.M.; Wan Zainon, W.M.N. Review of Feature Selection, Dimensionality Reduction and Classification for Chronic Disease Diagnosis. IEEE Access 2021, 9, 87310–87317. [Google Scholar] [CrossRef]
  24. Gupta, A.; Gupta, H.P.; Biswas, B.; Dutta, T. Approaches and Applications of Early Classification of Time Series: A Review. IEEE Trans. Artif. Intell. 2020, 1, 47–61. [Google Scholar] [CrossRef]
  25. Chamola, V.; Hassija, V.; Gupta, S.; Goyal, A.; Guizani, M.; Sikdar, B. Disaster and Pandemic Management Using Machine Learning: A Survey. IEEE Internet Things J. 2021, 8, 16047–16071. [Google Scholar] [CrossRef] [PubMed]
  26. Taherkhani, A.; Belatreche, A.; Li, Y.; Cosma, G.; Maguire, L.P.; McGinnity, T.M. A review of learning in biologically plausible spiking neural networks. Neural Netw. 2020, 122, 253–272. [Google Scholar] [CrossRef] [PubMed]
  27. Javanshir, A.; Nguyen, T.T.; Mahmud, M.A.P.; Kouzani, A.Z. Advancements in Algorithms and Neuromorphic Hardware for Spiking Neural Networks. Neural Comput. 2022, 34, 1289–1328. [Google Scholar] [CrossRef] [PubMed]
  28. Faraz, S.; Mellal, I.; Lankarany, M. Impact of Synaptic Strength on Propagation of Asynchronous Spikes in Biologically Realistic Feed-Forward Neural Network. IEEE J. Sel. Top. Signal Process. 2020, 14, 646–653. [Google Scholar] [CrossRef]
  29. Zahra, O.; Tolu, S.; Navarro-Alarcon, D. Differential mapping spiking neural network for sensor-based robot control. Bioinspiration Biomimetics 2021, 16, 036008. [Google Scholar] [CrossRef]
  30. Liu, C.; Wang, H.; Liu, N.; Yuan, Z. Optimizing the Neural Structure and Hyperparameters of Liquid State Machines Based on Evolutionary Membrane Algorithm. Mathematics 2022, 10, 1844. [Google Scholar] [CrossRef]
  31. Soffer, T.; Cohen, A. Students’ engagement characteristics predict success and completion of online courses. J. Comput. Assist. Learn. 2019, 35, 378–389. [Google Scholar] [CrossRef]
  32. Roschelle, J.; Feng, M.; Murphy, R.F.; Mason, C.A. Online Mathematics Homework Increases Student Achievement. Aera Open 2016, 2, 2332858416673968. [Google Scholar] [CrossRef]
  33. Muenks, K.; Wigfield, A.; Yang, J.S.; O’Neal, C.R. How True Is Grit? Assessing Its Relations to High School and College Students’ Personality Characteristics, Self-Regulation, Engagement, and Achievement. J. Educ. Psychol. 2017, 109, 599–620. [Google Scholar] [CrossRef]
  34. Hooda, M.; Rana, C.; Dahiya, O.; Rizwan, A.; Hossain, M.S. Artificial Intelligence for Assessment and Feedback to Enhance Student Success in Higher Education. Math. Probl. Eng. 2022, 2022, 5215722. [Google Scholar] [CrossRef]
  35. Comsa, I.M.; Potempa, K.; Versari, L.; Fischbacher, T.; Gesmundo, A.; Alakuijala, J. Temporal Coding in Spiking Neural Networks With Alpha Synaptic Function: Learning With Backpropagation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 2022, 5215722. [Google Scholar] [CrossRef] [PubMed]
  36. Kim, H.J.; Hong, A.J.; Song, H.D. The roles of academic engagement and digital readiness in students’ achievements in university e-learning environments. Int. J. Educ. Technol. High. Educ. 2019, 16, 1–18. [Google Scholar] [CrossRef]
  37. Friedrich, A.; Flunger, B.; Nagengast, B.; Jonkmann, K.; Trautwein, U. Pygmalion effects in the classroom: Teacher expectancy effects on students’ math achievement. Contemp. Educ. Psychol. 2015, 41, 1–12. [Google Scholar] [CrossRef]
  38. Aldowah, H.; Al-Samarraie, H.; Fauzy, W.M. Educational data mining and learning analytics for 21st century higher education: A review and synthesis. Telemat. Inform. 2019, 37, 13–49. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.