Next Article in Journal
Deep Learning for Predicting Congestive Heart Failure
Previous Article in Journal
Macro Model for Discrete-Time Sigma‒Delta Modulators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on Score Prediction Model with High Efficiency Based on Deep Learning

1
Department of Marxism, Hebei Chemical & Pharmaceutical College, Shijiazhuang 050026, China
2
School of Sciences, Hebei University of Science & Technology, Shijiazhuang 050018, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(23), 3995; https://doi.org/10.3390/electronics11233995
Submission received: 4 November 2022 / Revised: 22 November 2022 / Accepted: 25 November 2022 / Published: 2 December 2022
(This article belongs to the Section Artificial Intelligence)

Abstract

:
In the problem of unified classroom performance prediction, there is a certain lag in the prediction, and there are also problems such as the data sparsity and single feature in the data. In addition, feature engineering is often carried out manually in modeling, which highly depends on the professional knowledge and experience of engineers and affects the accuracy of the prediction to a certain extent. To solve the abovementioned gaps, we proposed an online course score prediction model with a high time efficiency that combines multiple features. The model uses a deep neural network, which can automatically carry out feature engineering and reduce the intervention of artificial feature engineering, thus significantly improving the time efficiency. Secondly, the model uses a factorization machine and two kinds of neural networks to consider the influence of first-order features, second-order features, and higher-order features at the same time, and it fully learns the relationship between the features and scores, which improves the prediction effect of the model compared to using only single feature learning. The performance of the model is evaluated on the learning analysis dataset from Fall 2015 to Spring 2021 and includes 412 courses with 600 students. The experimental results show that the performance of the prediction model based on the feature combination proposed in the present study is better than the previous performance prediction model. More importantly, our model has the best time efficiency of below 0.3 compared to the other models.

1. Introduction

In the information age, all walks of life have accumulated a large amount of data, but there is often some useful knowledge and valuable information hidden in the large amount of data. Machine learning and data mining technology can reveal some laws related to data and extract valuable information and data from them, which can be used to solve problems in various fields and provide help for administrators to make more reasonable and effective decisions. At present, machine learning and data mining-related technologies are widely used in the business, finance, medical, and other fields. In conclusion, in the context of education big data, the research on student performance prediction based on deep learning has important scientific significance and application prospects in promoting accurate management, scientific decision making, and improving education quality.
The quality of education is still the focus in the field of education, and improving the quality of education and teaching has always been one of the goals pursued by educators [1]. In the current educational environment, the evaluation of students learning effects is an issue that needs to be solved urgently [2]. In education, student achievement is an important indicator for evaluating the student development potential, development level, and student performance [2]. Student achievement prediction is a process of evaluating and inferring student information in a learning stage by mining and analyzing data based on curriculum settings, student historical grades, and student behavior to discover potentially effective information in teaching data. The predictions include course grades, Grade Point Average (GPA), course failure risk, and dropout risk [3]. Teachers or administrators can develop or refine teaching strategies, optimize resource allocation, and improve education and educational outcomes based on student abilities. Actions can also be taken based on the expected results. Recently, deep learning technology has made outstanding contributions in all walks of life, especially in the fields of unmanned aerial vehicle swarms [4], real-time systems [5], natural language processing, speech recognition, computer vision, etc. [6,7]. The value obtained by the wide application has also been valued by more and more researchers. Deep learning technology comes from the simulation of information transmission between neurons in the human brain. It extracts basic features from the input, further extracts multi-layer complex features, and learns more feature information. Compared with traditional data mining methods, deep learning algorithms rely on less manual feature extraction and are flexible. Therefore, the application of deep learning technology in student achievement prediction has become one of the important research topics [8].
At present, many scholars have conducted research on the prediction of student academic performance. In the early stage, students’ learning data (such as traditional classroom teaching test scores) were mainly collected from the educational administration system, and students’ consumption behavior data were collected from the all-in-one card to predict student performance [9]. Some researchers used the learner’s interactive records on the platform in the first week and after-school homework performance data, predicting whether a learner will eventually obtain a certificate based on a logistic regression model [10,11]. Brinton and MChiang developed an algorithmic model using factorization machines and K-Nearest Neighbors (KNN) to predict whether a student answered correctly the first time in a massive open online course (MOOC) question [12]. Lorenzo and Gomez-Sanchez utilized logistic regression, stochastic gradient descent, random forest, and support vector machine models to predict whether three engagement metrics (videos, exercises, and assignments) will decline relative to the previous metric at the end of the chapter [13]. Hlosta built a student performance prediction model based on machine learning methods (logistic regression, support vector machine, random forest, naive Bayes, and ensemble learning XGBoost) based on the data generated in the current course to assess whether students are at risk of dropping out of classes [14].
At present, the research on performance prediction has achieved certain results, but there are still some problems and deficiencies. First, in the traditional classroom grade prediction study, the data mainly come from some data generated during the course, such as the course assignment grades, unit test grades, etc. [15]. Therefore, the characteristics required for course grade prediction should be at the end of the course. The prediction results of the final grades are obtained late, the method has a certain lag, and the data also have problems such as sparse data and single features, so it cannot provide effective technical support for the teaching and management work in the early stage of the course [16]. Second, the existing online platform course grade prediction research mainly uses the log data of learners on the learning platform, such as the learners learning time on the online learning platform and the number of clicks on the learning video, and it lacks other relevant course grade information. In addition, the existing research often uses manual feature engineering in the modeling, which is highly dependent on the professional knowledge and experience of the engineers and affects the prediction accuracy of the method to a certain extent [17]. Finally, most of the data used in the existing performance prediction research come from the dataset constructed by the researchers themselves, and the amount of data is generally small. For mainstream research methods such as machine learning algorithms, there are certain requirements for the amount of data. If the amount of data is insufficient, it is difficult to train a better model, which leads to a low prediction accuracy to a certain extent [18].
Educational data mining techniques can provide educational decision makers with data-based models that are essential to support their goals of improving the efficiency and quality of teaching and learning. The main applications of educational data mining are student performance prediction, the discovery of student bad behavior, knowledge tracking, course recommendation, etc. [19]. Among them, student achievement prediction is a typical problem in educational data mining. Predicting student achievement can help students with learning difficulties in time, and at the same time, it provides teaching suggestions and theoretical support for teaching managers and teachers [20]. Deep learning is a neural network structure based on multiple layers of processing units, which is used in wide fields of practical applications, such as automatic driving [21], path planning [22], natural language processing, and speech recognition. The spread of the deep learning technology allowed college students to create early warning models.
Aiming at the problems of the sparse and lagging prediction data of a traditional classroom in student scores in education data mining, this paper proposed a traditional classroom score prediction model that integrates a self-attention mechanism and depth matrix decomposition, using the course scores learned in previous semesters to predict the course scores to be learned in the next semester or several semesters. First, the self-attention mechanism is added to the model, which can quickly extract the important potential features of the students and courses and make the model more focused on useful information. Secondly, a bilinear pooling layer is built in the model to improve the generalization and learning ability of the model.
As for the problem of online course student performance prediction, the achievement of student performance will be affected by many factors, aiming at the problem that the existing in-depth learning methods of performance prediction do not consider the impact of multiple features on performance prediction at the same time. Therefore, this paper proposed an online course performance prediction model that combines multiple features. Firstly, the model can automatically carry out feature engineering by using a deep neural network, which reduces the intervention of artificial feature engineering. Secondly, the model uses a factorization machine and two kinds of neural networks to consider the influence of first-order features, second-order features, and higher-order features at the same time, fully learning the relationship between the features and grades and improving the prediction effect of the model compared to using only single feature learning.
Our paper is structured as follows. In the Methods and data Section 2, we introduce the method in terms of the deep matrix factorization, dataset, and prediction model of online course grades based on feature combination. The main results of this study are organized in Section 3. Finally, we conclude the main idea in Section 4. Our study highlights that (1) the student achievement prediction model, which combines the self-attention mechanism and depth matrix decomposition, can overcome the problem of lagging prediction results in traditional classroom achievement prediction, and (2) the new feature combination structure model in the present study can overcome the shortcomings of the existing online course score prediction methods.

2. Methods and Data

This paper proposed a method to predict student performance, which integrates self-attention mechanism and depth matrix decomposition. Applying the deep matrix decomposition model integrating the self-attention mechanism to predict the course scores that students will learn next semester, similar to the matrix decomposition model.

2.1. Deep Matrix Factorization

Recently, however, it was found to be applied to the prediction of student grades. For example, the problem of the prediction of students’ curriculum grades in the next semester is analogized to the problem of score prediction or that of the next basket recommendation [23,24,25]. The method of the recommendation system has brought a new perspective to solve the problem of student grades prediction. In particular, the matrix decomposition method decomposes the student curriculum score matrix into two low-rank matrices to represent the curriculum and student potential knowledge space, and then obtains the scores of the unfinished courses through the dot product of the corresponding vectors in the two decomposition matrices [26]. The depth matrix decomposition model was successfully applied in the field of recommendation systems. For example, Xue et al. proposed a depth matrix decomposition model to decompose the scoring matrix of users and items and then mapped the decomposed two potential feature vectors to a low-dimensional space through two sets of multi-layer perceptrons, finally realizing the user scoring prediction of movies [27]. Therefore, it can also be used to predict student curriculum achievements.
Assuming m represents number of students and n is the courses number in Tth semesters, a matrix G R m × n is student–course grades, each item in the G includes a tuple for the student and course. Therefore, each value in G is denoted as g(s, c), which represents the achievement g(s, c) of student s in course c, g ( s , c ) [ 0 , 5 ] . In the T semester, the student–course grade (SCG) is expressed as G T = I = 1 T G i .
A grade matrix G is constructed by T semester student–course data in the input layer, where G(i) is the grades achieved by students in each course, the Gj is the grades of each student in each course,   ( i , j ) Ω . Because the constructed grade matrix is very sparse, the data dimension can be reduced, and the data sparsity can be alleviated through the embedding-layer mapping [28]. For a given score matrix Gij, forms of grade matrix are obtained as follows:
x s = T a n h ( G i · W s ) ,   x c = T a n h ( G j · W c )
where x s ,   x c G k is student latent eigenvectors and course latent eigenvectors, respectively.   W s G m × k , W c G k × n is the feature projection weight matrix. In the self-attention layer, the student LFV (latent feature vector) and the course LFV are added with different weights. Varied student features and course features have varying importance. The impact of marketing courses on the advanced computer network courses to be studied in the next semester is different. A computer network is the basic subject of advanced computer network, so the relative weight of its influence will be larger, and it will be transformed according to the following formula:
x s a = s o f t m a x ( x s · W s a ) · x s ,   x c a = s o f t m a x ( x c · W c a ) · x c
where W s a ,   W c a   are self-attention distribution weight parameters. x s a , x s c is the student latent eigenvectors and course latent eigenvectors with weight values, respectively. The multi-layer perceptron layer, it is transformed according to Equation (3), and it aimed to better learn the nonlinear characteristics of the information and perform a two-layer nonlinear mapping between the latent options of the scholars and course.
{ H s = s i g m o i d ( x s a · W s + b s ) p s = s i g m o i d ( H s · W s + b s ) H c = s i g m o i d ( H c · W c + b c ) q c = s i g m o i d ( H c · W c + b c )
The following transformations are performed in the bilinear pooling layer.
g ¯ = h T σ ( p s T q c )
where h, , σ is weight parameter, Hadamard product, activation function, respectively.
The model that predicts the course is built by following:
L = m i n W , W S , W C , h 1 2 i , j W ( G i j h T σ ( ( p s ( i ) ) T q c j ) ) 2 + λ ( | | W s | | 2 F + | | W c | | F 2 + | | W | | 2 )      
where unknown parameter set, h T σ ( ( p s ( i ) ) T q c j ) ) 2 , is the predicted value of the model output through the nonlinear pooling layer. In addition, learn model parameters with Adams optimization method, using Sigmoid as an activation function.

2.2. Dataset

The data used in our work collect all grades from Fall 2015 to Spring 2021 and include 412 courses with 600 students. In the dataset, students course scores range from 0 to 100. Because the predicted target values in the model proposed in present study are discrete, the 0–59 points are converted to 1, the 60–69 points are converted to 2, the 70–79 points are converted to 3, and the 80–89 points are converted to 4. Convert 90–100 points to 5. Train the model on the (T − 1)th semester dataset and predict the course grades for the Tth semester. Use the dataset from Autumn 2015 to Autumn 2020 to train the model, and test the model proposed in this chapter on the dataset in Spring 2021 (i.e., T = Spring 2021). The dataset is divided as follows:
Dataset 1: train data are from 2015 Fall to 2020 Fall, test data are 2021 Spring; dataset 2: the train data are from 2015 Fall to 2020 Spring, the test data are 2021 Fall; dataset 3: the train data are from 2015 Fall to 2019 Fall, the test data are 2019 fall.
The method of matrix decomposition can be applied. For example, a performance matrix is decomposed into 2 low-rank matrices containing latent factors of courses and students. p s G k and q c G k is students s and courses c latent k-dimensional features. Equation (6) can be used to predict the grade of student s in course c.
g s , c = p s T q c ^
The loss function of MF is:
L = m i n p , q s , c Ω | | g s , c p s q c T 1 | | 2 + 1 λ ( | | p s | | 2 + | | q c | | 2 )
where L2 is a regular term to prevent overfitting.

2.3. Prediction Model of Online Course Grades Based on Feature Combination

Internet, more and more online learning platforms have emerged, and student performance prediction based on online learning behavior data has become a development trend. For the prediction of student grades in online courses, the achievement of student grades will be affected by many factors. It is not as long as the study time is longer or the students who answer the questions better after class can get good grades. Feature combination refers to considering the influence of multiple features on the entire prediction problem at the same time in practical application scenarios.
We proposed a new online course grade prediction method, which is based on different structural combination features, and its main work has three aspects. (1) The study found that considering two or more features at the same time has a certain impact on the prediction results of student grades; (2) the proposed feature combination model can automatically learn the combination of features and consider the effects of first-, second-, and high-order features at the same time; (3) the performance of the model is verified on the open UK Open University learning analysis dataset, which has high accuracy and effectiveness.
Given a set of student features F, F consists of student attribute features student_attr = { s 1 , , s m } ,   course attributes course_attr = { c 1 , , c n } ,   b e h a v i o r = { b 1 , , b k } . Among them, m, n, k are the number of features, respectively. For the student, their final course grade is yi, y = {0, 1} is the set of categories divided by the student grade, where 0 means that the student has failed, and 1 means that the student has passed. The goal of student grade prediction is to predict the grade category yi obtained by students based on student characteristics F.
Our study aimed to mine data related to student learning based on deep learning technology, to achieve accurate prediction of student academic performance and to provide timely help and guidance to students who have “risks”. Therefore, this study proposes a feature combination-based performance prediction model (factorization deep product neural network, FDPN) based on student attribute characteristics, course attribute characteristics, and student learning behavior characteristics. The model framework is shown in Figure 1. FDPN consists of 3 layers. Embedding layer: reduce the scale of the initial high-dimensional options and map them to low-dimensional feature vectors. Network layer: this layer consists of factorization machine (FM), deep neural network (DNN), and product neural network (PNN). FM is used to learn the first-order and second-order feature representations of features, and DNN and PNN are, respectively, used to learn higher-order representations between features. Prediction layer: by splicing, the low- and high-order features are combined to obtain the final features with richer information so as to make better student performance predictions.
In this paper, the cross entropy loss function is used, and the L2 regularization parameter is used. The loss function of the model is as follows:
l o s s = 1 n i = 1 n y i log g + λ | | θ | | 2
Among them, n is the total number of training data, yi is the grade category of the i data, g is the predicted probability of the grade category of the i data, λ | | θ | | 2 is L2 regular term, θ is the set of all parameters of the model.
The evaluation indicators in this paper use accuracy, precision, recall, F1, and AUC (Area Under Curve) to measure the performance of model classification and prediction.
{ A c c u r a c y = T P + T N T P + T N + F P + F N P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
This paper uses the Open University Learning Analysis Dataset (OULAD), which contains multiple data types, such as the basic information from the Open University in the UK from 2015 to 2021. First, the Open University opens a course, students apply for course registration, and then the course can start to study. The course duration of the Open University is mostly nine months. During the learning process, there will be corresponding learning tasks that the learners must complete.
Three widely used evaluation indicators are used to evaluate the performance of the proposed model.
R M S E = 1 Ω i , j Ω ( y i j y i , j ^ ) 2
M A E = 1 Ω i , j Ω y i j y i j ^
M A P E = 100 Ω i = 1 Ω y i j y i j ^ y i
where y i j is observations, y i j ^ is average value of observations. RMSE is root mean standard error, MAE is mean absolute error, MAPE is mean absolute percent error. RM measures the deviation between the observed value and the true value. MAE can better reflect the actual situation of prediction error. MAPE can measure the advantages and disadvantages of the model. An MAPE of 0% indicates a perfect model, while an MAPE greater than 100% indicates a poor quality model.

3. Results and Discussion

This experiment trains attention deep matrix factorization (ADFM) models on three datasets. Figure 2 shows the results of the experiments with the ADMF models on three datasets. The ADFM model has the smallest root mean square error (RMSE), mean absolute error (MAE), and mean absolute percent error (MAPE) on dataset 1, and the model prediction effect is the best. In this experiment, the amount of coaching knowledge in the dataset, a pair of, and dataset three is considerably below that in dataset one. Therefore, we will carry out comparative experiments on the dataset.
From the comparison of the experimental results of the DMF and ADMF in Figure 3, which show that the RMSE of the ADMF model is 84%, the MAE is 66.7%, and the MAPE is 21%, which is smaller than that of the DMF model. Therefore, incorporating a self-attention mechanism into a deep neural network can permit the model to pay much attention to the knowledge which is important for prediction accuracy. During this experiment, the potential feature of the course and students have different weights and different influences on the student performance; therefore, it is clear that the ADMF model contains a sensible impact. The effect of the bilinear pooling layers on the model performance is shown in Table 1.
We compare two ways to obtain the predicted values, bilinear pooling, and inner product. From the results in Table 1, it is shown that the prediction effect of the bilinear pooling mapping is better than that of the inner product model. The performance of the prediction model proposed in the present study uses the bilinear pooling method to archive the output, which can effectively improve the performance of the model. Regarding the generalization ability, a better performance prediction effect can be obtained.
Several traditional methods of predicting student performance were compared, including the KNN, MF, NCF, and DMF. They were tested on dataset 1, and the final experimental results are shown in Figure 4. We found that the NCF, DMF, and ADMF accomplish higher prediction results than the MF and KNN. The ADMF is the best predictor for students in the next semester. Because the MF is extended to multi-layer neural networks, the ADMF encompasses a stronger ability to spot the input file, which will learn complicated nonlinear structural options hidden in the information, and might conjointly mix easy options into much more complicated options, with flexibility and accuracy. This method decomposes the student course score matrix into row vectors and column vectors, which represent the student scores of each course and all student scores of a specific course, respectively. Then, the row vectors and column vectors are embedded into a low-dimensional space and input into a multi-layer perceptron network based on a self-attention mechanism to further extract information from potential feature vectors, in which the attention mechanism is added, and it can quickly extract the important potential features of the students and courses. Secondly, a bilinear pooling layer is built in the model to improve the generalization and learning ability of the model.
Our model has two deep neural networks. When the network contains multiple hidden layers, if the number of neurons in the hidden layers is different, many experiments are required. The number of neurons in the layer is set to the same number. The experimental results are shown in Figure 5. This paper carried out six experiments, changing the number of neurons in turn. From the experimental results, when the number of neurons is 256, the recall rate is about 95%, the precision is about 90.7%, the AUC is about 82%, the accuracy rate is about 86.6%, and its model performance is the best. This is mainly because as the number of neurons increases, the model can learn more feature information, but once the variety of neurons increases to a precise number, the model learns the effective information is not increasing, and it will even introduce noise that reduces the prediction performance of the model. It is seen from this that once it involves deep neural network coaching, the quantity of neurons is not the maximum amount as the potential, neither is it as little as the potential. It is necessary to incessantly compare the model coaching and learn to pick out the optimum range of neurons.
In our study, we also compared the activation function of the hidden layer neurons with the ReLU activation function and Tanh activation function. The prediction effect of the model is related to the activation function setting of the hidden neurons. Because this is a binary model, the final prediction output unit of the model uses the sigmoid function, and the other activation functions are set the same. The experimental results are shown in Table 2. It can be seen from the table that when the activation function is the ReLU, the prediction accuracy, recall rate, F1, and AUC of the FDPN model are increased by about 2%, and the prediction effect of the hidden layer neuron activation function using ReLU is better than that of the Tanh function. In the feedforward neural network, the performance of the ReLU activation function will be better than that of the Tanh performance, and the activation performance continues to be set as ReLU within the ensuant experiments of this paper.
The model proposed in this paper includes two feedforward neural networks, the DNN and PNN. The number of hidden layers in the neural network is different, and the prediction ability of the model is different. To simplify the experiment, this paper sets the same number of hidden layers and neurons in each layer of the two neural networks, that is, the number of hidden layers is increased from one to three. The experimental results are shown in Table 3. It can be seen from the experimental results that when the number of hidden layers is one and the number of hidden layers is two, the model performance is better, and when the number of hidden layers is three, the prediction effect of the model decreases significantly. As the number of layers increases, each evaluation index decreases, mainly because the more layers, the more complex the structure, the greater the amount of calculation, and the problem of the overfitting of the model will occur. Therefore, the number of hidden layers is still set to one in the subsequent experiments of this paper.
The FM learns the first-order and second-order representations of the features, and the DNN and PNN learn different high-order representations of features. In this paper, three structures are combined to learn to predict grades. In the experiments, different feature combination structures are compared to investigate the impact of the structure on the prediction skill of the model. For the results, see Table 4.
The experimental results show that the result of the one-structure FM, DNN, and PNN is worse, the combined feature learning of the DeepFM, DNN + PNN, and FM + PNN of the two structures is higher than that of the one structure, and therefore the combined feature learning result of the three structures of the FDPN model is the best. This is because the FDPN model simultaneously considers the first-order feature representation, the second-order feature representation, and two different high-order feature representations and utilizes more potentially effective information in the grade prediction. To sum up, the FDPN performance prediction model incorporates a smart prediction impact and might improve the performance prediction performance.
In this paper, we use the LR, SVM, FM, and deep learning models such as the DNN, DeepFM, and PNN as the comparison models. Through the comparison experiments on the OULAD dataset, the results are shown in Table 5, which verifies that the FDPN model proposed in this paper has the best prediction. From the experimental results in Table 5, it can be seen that compared with the existing performance prediction methods (LR, SVM, FM, and DNN), the method based on feature combination on the OULAD public education dataset has achieved the best prediction effect. Compared with the best traditional method, the DNN accuracy and AUC are both improved by 2%. In addition, there are also vital enhancements within the three indicators of exactness, recall, and F1. The tactic that supported the feature combination projected during this paper is healthier than the four ancient performance prediction ways within the comparative experiment, mainly as a result of the normal performance prediction methodology which uses every attribute feature directly as a classification feature input model for learning and coaching, and only considers low-level features or high-order features, without considering the different effects of low- and high-order feature combinations on the final grades. For the other two feature combination methods (DeepFM and PNN) in the comparative experiment, the method proposed in this paper extracts more feature information for each attribute feature, including first-order features, second-order features, and two different high-order features. Thereby, the prediction ability of the model is greatly improved, the prediction effect is better, and the validity of the model is finally proved through the experiments.

4. Conclusions

(1) This paper proposed a student performance prediction model that integrates a self-attention mechanism and depth matrix decomposition. The experimental results show that the method proposed in this chapter is superior to the benchmark comparison method. The RMSE of the ADMF model is 84%, the MAE is 66.7%, and the MAPE is 21%, which is smaller than that of the DMF model.
(2) A new feature combination structure model is proposed to overcome the shortcomings of the existing online course score prediction methods. The experimental results show that the model proposed in this study has a good performance prediction ability. When the number of neurons is 256, the recall rate is about 95%, the precision is about 90.7%, the AUC is about 82%, and the accuracy rate is about 86.6%.
However, it does not consider the impact of the time series of students learning courses, student comments on online platforms, and the other contents on student scores. (i) Because of the traditional classroom performance prediction problem, this paper only uses the final scores of multiple courses to collect more student-related data, such as the number of times students borrow books every week, relevant information about teachers, etc., considering the impact of multiple characteristics on the student curriculum performance, and further considering the relationship between courses. (ii) For the online platform course score prediction problem, we can consider more in-depth learning algorithm models, consider the different effects of different features, and use the attention mechanism to learn their different weights to further improve the prediction performance of the model. (iii) For the above two aspects of the performance prediction, the time series of course learning has not been considered. In future work, we can consider the influence of this factor to further explore.

Author Contributions

Conceptualization, methodology, validation, formal analysis, investigation, resources and writing, L.Y.; data curation, writing—original draft preparation, review and editing, Z.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work. There is no professional or other personal interest of any nature or kind in any product, service, and/or company that could be construed as influencing the position presented in, or the review of, the manuscript entitled.

References

  1. Romero, C.; Ventura, S. Educational data mining: A review of the state of the art. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2010, 40, 601–618. [Google Scholar] [CrossRef]
  2. Ortiz-Echeverri, C.J.; Rodríguez-Reséndiz, J.; Garduño-Aparicio, M. An approach to STFT and CWT learning through music hands-on labs. Comput. Appl. Eng. Educ. 2018, 26, 2026–2035. [Google Scholar] [CrossRef]
  3. Ning, X.; Tian, W.; Yu, Z.; Li, W.; Bai, X.; Wang, Y. HCFNN: High-order Coverage Function Neural Network for Image Classification. Pattern Recognit. 2022, 131, 108873. [Google Scholar] [CrossRef]
  4. Aher, S.B.; Lobo, L. Combination of machine learning algorithms for recommendation of courses in E-Learning System based on historical data. Knowl. -Based Syst. 2013, 51, 1–14. [Google Scholar] [CrossRef]
  5. Ezz, M.; Elshenawy, A. Adaptive recommendation system using machine learning algorithms for predicting students best academic program. Educ. Inf. Technol. 2020, 25, 2733–2746. [Google Scholar] [CrossRef]
  6. Zhou, Y.; Huang, C.; Hu, Q.; Zhu, J.; Tang, Y. Personalized learning full-path recommendation model based on LSTM neural networks. Inf. Sci. 2018, 444, 135–152. [Google Scholar] [CrossRef]
  7. Ning, X.; Xu, S.; Nan, F.; Zeng, Q.; Wang, C.; Cai, W.; Jiang, Y. Face editing based on facial recognition features. IEEE Trans. Cogn. Dev. Syst. 2022, 1–2. [Google Scholar] [CrossRef]
  8. Mota-Valtierra, G.; Rodríguez-Reséndiz, J.; Herrera-Ruiz, G. Constructivism-Based Methodology for Teaching Artificial Intelligence Topics Focused on Sustainable Development. Sustainability 2019, 11, 4642. [Google Scholar] [CrossRef] [Green Version]
  9. Ucar, H.; Kumtepe, A.T. Effects of the ARCS-V-based motivational strategies on online learners’ academic performance, motivation, volition, and course interest. J. Comput. Assist. Learn. 2020, 36, 335–349. [Google Scholar] [CrossRef]
  10. Ren, Z.; Rangwala, H.; Johri, A. Predicting performance on MOOC assessments using multi-regression models. arXiv 2016, arXiv:1605.02269. [Google Scholar]
  11. Sweeney, M.; Rangwala, H.; Lester, J.; Johri, A. Next-term student performance prediction: A recommender systems approach. arXiv 2016, arXiv:1604.01840. [Google Scholar]
  12. Polyzou, A.; Karypis, G. Grade prediction with models specific to students and courses. Int. J. Data Sci. Anal. 2016, 2, 159–171. [Google Scholar] [CrossRef]
  13. Aljohani, N.R.; Fayoumi, A.; Hassan, S.U. Predicting at-risk students using clickstream data in the virtual learning environment. Sustainability 2019, 11, 7238. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, J.; Zhang, Y.; Wu, L.; You, T.; Ning, X. An adaptive clustering-based algorithm for automatic path planning of heterogeneous UAVs. IEEE Trans. Intell. Transp. Syst. 2021, 23, 16842–16853. [Google Scholar] [CrossRef]
  15. Chen, J.; Ling, F.; Zhang, Y.; You, T.; Liu, Y.; Du, X. Coverage path planning of heterogeneous unmanned aerial vehicles based on ant colony system. Swarm Evol. Comput. 2022, 69, 101005. [Google Scholar] [CrossRef]
  16. Aldowah, H.; Al-Samarraie, H.; Fauzy, W.M. Educational data mining and learning analytics for 21st century higher education: A review and synthesis. Telemat. Inform. 2019, 37, 13–49. [Google Scholar] [CrossRef]
  17. Du, X.; Liu, Y.; Lu, Z.; Duan, Q.; Feng, J.; Wu, J.; Chen, B.; Zheng, Q. A Low-Latency Communication Design for Brain Simulations. IEEE Netw. 2022, 36, 8–15. [Google Scholar] [CrossRef]
  18. Pero, Š.; Horváth, T. Comparison of collaborative-filtering techniques for small-scale student performance prediction task. In Innovations and Advances in Computing, Informatics, Systems Sciences, Networking and Engineering; Springer: Cham, Switzerland, 2015; pp. 111–116. [Google Scholar]
  19. Tomasevic, N.; Gvozdenovic, N.; Vranes, S. An overview and comparison of supervised data mining techniques for student exam performance prediction. Comput. Educ. 2020, 143, 103676. [Google Scholar] [CrossRef]
  20. Waheed, H.; Hassan, S.U.; Aljohani, N.R.; Hardman, J.; Alelyani, S.; Nawaz, R. Predicting academic performance of students from VLE big data using deep learning models. Comput. Hum. Behav. 2020, 104, 106189. [Google Scholar] [CrossRef] [Green Version]
  21. Dai, C.; Liu, X.; Lai, J.; Li, P.; Chao, H.C. Human behavior deep recognition architecture for smart city applications in the 5G environment. IEEE Netw. 2019, 33, 206–211. [Google Scholar] [CrossRef]
  22. Guo, Z.; Liu, T.; Chang, L.; Gu, T. Deep matrix factorization based on self-attention mechanism for student grade prediction. J. Phys. Conf. Ser. 2020, 1651, 012019. [Google Scholar] [CrossRef]
  23. Jin, X.; Miao, J.; Wang, Q.; Geng, G.; Huang, K. Sparse matrix factorization with [formula omitted] norm for matrix completion. Pattern Recognit. 2022, 127, 108655. [Google Scholar] [CrossRef]
  24. Duan, R.; Jiang, C.; Jain, H.K. Combining review-based collaborative filtering and matrix factorization: A solution to ratings sparsity problem. Decis. Support Syst. 2022, 156, 113748. [Google Scholar] [CrossRef]
  25. Vangala, J. Parallel Non-Negative Matrix Factorization: Gene Expression Analysis. J. Comput. Sci. Syst. Biol. 2021, 14, 380. [Google Scholar]
  26. Dou, Y.; Yun, Y.; Gao, Q.; Zhang, X. Self-representation and matrix factorization based multi-view clustering. Neurocomputing 2021, 459, 395–407. [Google Scholar] [CrossRef]
  27. Wang, C.; Ning, X.; Sun, L.; Zhang, L.; Li, W.; Bai, X. Learning Discriminative Features by Covering Local Geometric Space for Point Cloud Analysis. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5703215. [Google Scholar] [CrossRef]
  28. Skytt, V.; Kermarrec, G.; Dokken, T. LR B-splines to approximate bathymetry datasets: An improved statistical criterion to judge the goodness of fit. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102894. [Google Scholar] [CrossRef]
  29. Maya, S.; Richard, J.; Mehdi, S.; Kristen, F.; Stanislav, H. SVM Communications: Supervised exercise therapy for symptomatic peripheral artery disease—A conversation with the experts. Vasc. Med. 2022, 27, 214–216. [Google Scholar]
  30. Zhou, J.; Li, X.; Wang, X.; Chai, Y.; Zhang, Q. Locally weighted factorization machine with fuzzy partition for elderly readmission prediction. Knowl. -Based Syst. 2022, 242, 108326. [Google Scholar] [CrossRef]
  31. Wei, Z.; Han, X.; Li, J. State of health assessment for echelon utilization batteries based on deep neural network learning with error correction. J. Energy Storage 2022, 51, 104428. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Wu, L.; Deng, L.; Ouyang, B. Retrieval of water quality parameters from hyperspectral images using a hybrid feedback deep factorization machine model. Water Res. 2021, 204, 117618. [Google Scholar] [CrossRef] [PubMed]
  33. Yaqub, M.; Lee, S.H. Experimental and neural network modeling of micellar enhanced ultrafiltration for arsenic removal from aqueous solution. Environ. Eng. Res. 2021, 26, 190261. [Google Scholar] [CrossRef]
Figure 1. Framework of FDPN model. Factorization machine (FM); deep neural network (DNN); product neural network (PNN).
Figure 1. Framework of FDPN model. Factorization machine (FM); deep neural network (DNN); product neural network (PNN).
Electronics 11 03995 g001
Figure 2. DMF prediction effect on different datasets. The description of dataset is introduced in Section 2.2.
Figure 2. DMF prediction effect on different datasets. The description of dataset is introduced in Section 2.2.
Electronics 11 03995 g002
Figure 3. Comparison of experimental results between DMF and ADMF.
Figure 3. Comparison of experimental results between DMF and ADMF.
Electronics 11 03995 g003
Figure 4. Different models predict the effect. Time efficiency is defined by ratio of time consuming for one model to total time for total models.
Figure 4. Different models predict the effect. Time efficiency is defined by ratio of time consuming for one model to total time for total models.
Electronics 11 03995 g004
Figure 5. Comparison of the number of neural units in the hidden layer.
Figure 5. Comparison of the number of neural units in the hidden layer.
Electronics 11 03995 g005
Table 1. Comparison of predicted value output methods.
Table 1. Comparison of predicted value output methods.
Predicted Value Output MethodRMSEMAEMAPE
Bilinear pooling0.84020.65230.2315
Inner product0.84780.67890.2355
Table 2. Comparison of activation function.
Table 2. Comparison of activation function.
Activation FunctionAccuracyRecallF1AUCPrecision
Tanh0.84450.86100.92090.79880.8610
ReLU0.86690.86800.94090.81850.8680
Table 3. Hidden layer comparison.
Table 3. Hidden layer comparison.
LayersAccuracyPrecisionRecallF1AUC
10.86680.86890.95880.90120.8125
20.85020.86320.92130.89520.8050
30.78560.88120.78520.83650.7788
Table 4. Structures comparison.
Table 4. Structures comparison.
StructureAccuracyPrecisionRecallF1AUC
FM0.8410.85220.92790.89000.7920
DNN0.84390.85490.92900.89100.7940
DeepFM0.84830.86000.93970.89730.8010
PNN0.85860.84640.93980.80040.8100
DNN + PNN0.86410.86600.94250.89370.8131
FM + PNN0.8250.86700.94400.90320.8174
FDPN0.86690.86890.94990.90710.8179
Table 5. Model comparison. Efficiency is defined by 1-(ratio of time consuming for one model to total time for all models.
Table 5. Model comparison. Efficiency is defined by 1-(ratio of time consuming for one model to total time for all models.
ModelAccuracyPrecisionRecallF1EfficiencyAUC
LR [28]0.79890.83390.87260.85590.30.7552
SVM [29]0.83200.84320.92250.88290.40.7789
FM [30]0.84290.84650.92790.89000.560.7939
DNN [31]0.84390.85520.92900.89100.610.8008
DeepFM [32]0.84800.86230.92890.89400.550.8002
PNN [33]0.85460.86560.93890.90020.750.8089
FDPN0.86580.86790.94980.90690.860.8180
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, L.; Bai, Z. Study on Score Prediction Model with High Efficiency Based on Deep Learning. Electronics 2022, 11, 3995. https://doi.org/10.3390/electronics11233995

AMA Style

Yang L, Bai Z. Study on Score Prediction Model with High Efficiency Based on Deep Learning. Electronics. 2022; 11(23):3995. https://doi.org/10.3390/electronics11233995

Chicago/Turabian Style

Yang, Lihong, and Zhiming Bai. 2022. "Study on Score Prediction Model with High Efficiency Based on Deep Learning" Electronics 11, no. 23: 3995. https://doi.org/10.3390/electronics11233995

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop