Next Article in Journal
Entropy of the Multi-Channel EEG Recordings Identifies the Distributed Signatures of Negative, Neutral and Positive Affect in Whole-Brain Variability
Previous Article in Journal
Entropy Generation and Heat Transfer in Drilling Nanoliquids with Clay Nanoparticles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction

1
School of Information Science and Technology, Northwest University, Xi’an 710127, China
2
Maths and Information Technology School, Yuncheng University, Yuncheng 044000, China
3
State-Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services, Northwest University, Xi’an 710127, China
4
College of Economic and Management, Xi’an University of Posts and Telecommunications, Xi’an 710016, China
5
China Aerospace Academy of Systems Science and Engineering, Beijing 100048, China
*
Authors to whom correspondence should be addressed.
Entropy 2019, 21(12), 1227; https://doi.org/10.3390/e21121227
Submission received: 22 October 2019 / Revised: 5 December 2019 / Accepted: 13 December 2019 / Published: 16 December 2019

Abstract

:
Personnel performance is important for the high-technology industry to ensure its core competitive advantages are present. Therefore, predicting personnel performance is an important research area in human resource management (HRM). In this paper, to improve prediction performance, we propose a novel framework for personnel performance prediction to help decision-makers to forecast future personnel performance and recruit the best suitable talents. Firstly, a hybrid convolutional recurrent neural network (CRNN) model based on self-attention mechanism is presented, which can automatically learn discriminative features and capture global contextual information from personnel performance data. Moreover, we treat the prediction problem as a classification task. Then, the k-nearest neighbor (KNN) classifier was used to predict personnel performance. The proposed framework is applied to a real case of personnel performance prediction. The experimental results demonstrate that the presented approach achieves significant performance improvement for personnel performance compared to existing methods.

1. Introduction

Modern organizations are facing a higher level of increasingly complex competition in global markets. The future survival of organizations depends primarily on personnel performance. Personnel performance, such as capability, conscientiousness, achievement motivation, and other characteristics, plays a critical role in maintaining the competitive advantages of an organization. Accurate personnel performance prediction can help decision makers to recruit and select the most appropriate people for each job. Consequently, personnel performance predicted by computer-aided techniques is an extraordinary challenging research topic.
Many researchers pay close attention to the study on human performance modeling. In the literature, there are many efforts toward forecasting human performance using various methods from psychology, management and data mining areas, such as correlation analysis and rule-based methods [1]. However, rule-based methods have limited performance, because domain experts or professionals need to manually predesign large bodies of rules to explore the meaningful patterns to some extent. Recently, machine-learning-based methods have attracted great attention from some researchers. These methods have also been successfully applied to human performance modeling. Chien et al. [2] presented a data-mining framework using decision tree and association rules for human performance modeling. This study extracted useful patterns and rules between personnel characteristics and work behaviors. A self-regulating clustering algorithm was presented to analyze bank personnel performance in order to improve the performance of human resource management (HRM) [3]. Gobert et al. [4] developed an interactive environment to evaluate students’ scientific research capability via data mining technology. Li et al. [5] proposed an improved KNN algorithm to deal with human performance prediction in a manufacturing system. This method utilized a distance calculation formula based on entropy, a classification rule, and a quantitative description way of human performance. Wang et al. [6] applied a hybrid feature selection method to handle human resource selection, which increased the classification performance. Although many methods have been proposed, depending on the characteristics of good features, these traditional machine-learning-based methods have weak generalization and learning ability. The performance relies too much on prior knowledge of domain designers, which is cumbrous and time-consuming, and heavily affects classification accuracy. However, human performance modeling research is still at an early stage, and there is great potential for further improvements in its performance.
In recent years, deep learning [7,8] and artificial neural networks have exhibited outstanding performance, which have become the state-of-the-art method for many pattern recognition problems. They have been successfully applied in a range of computer tasks, including image classification [9], breast mass classification [10], semantic segmentation [11], object detection and recognition [12,13], emotion recognition [14], language identification [15], agricultural areas [16], drug–drug interaction extraction [17], and so on. However, few works can be found in the field of HRM.
Lately, the rapid development of deep learning has brought new inspiration to HRM tasks. Convolutional neural networks (CNN) [18], which have made a significant breakthrough in feature learning, are a well-known deep learning method. Rather than handcrafted features being extracted, CNN can automatically learn more discriminative features for the current classification task. However, CNN cannot capture the dependency between features which are some distance apart and have semantic relevance. Therefore, we propose a hybrid CRNN model combining CNN with long short-term memory (LSTM) to extract the target features of given personnel performance prediction samples. To better capture global personnel performance information, we introduced self-attention to our CRNN model.
Particularly in the current circumstances, there is generally no prior knowledge about data distribution characteristics. The KNN classifier is the most suitable for processing these data. In addition, KNN can acquire more complicated decision information than the general softmax/sigmoid activation functions at the last layer of self-attention-based CRNN. Self-attention-based CRNN and KNN are completely complementary in terms of feature extraction and decision information. Consequently, we put forward a novel Convolutional Hybrid Recurrent Neural Networks with a Self-Attention Mechanism (called CHRNNA) to make full use of their advantages and be able to make up for their deficiencies.
Recent advances in machine learning suggest that, given sufficient labeled data, there should be an opportunity to construct better prediction models. However, there is no manual labeling of data publicly available. In this work, we created a labeled personnel performance dataset and exploited it using deep learning methods to build an accurate model for personnel performance prediction.
Overall, our contributions in this paper are as follows:
(1)
Grounded in psychological theory and management theory of personnel performance, we constructed a high-quality dataset with performance prediction characteristics. This is key to ensuring data quantity and quality;
(2)
To the best of our knowledge, this is first time deep learning has been applied to the field of personnel performance prediction, which fills the gap in this field;
(3)
Instead of considering each attribute equally, a self-attention mechanism was used to automatically select the informative features of personnel performance. Our proposed CHRNNA framework can be viewed as a universal framework for personnel performance prediction, which has greatly improved classification performance.
The remainder of this paper is structured as follows: The related background is introduced in Section 2; data are characterized in detail, and our proposed method is described thoroughly in Section 3; Section 4 evaluates the effectiveness of our method in a wide range of experiments and presents experimental results as well as analysis; finally, we present general conclusions from our work in Section 5.

2. Related Work

The related work can be classified into the following categories, i.e., personnel performance prediction and deep learning research.

2.1. Personnel Performance Prediction

In recent years, more and more research results have been devoted to personnel performance prediction. Existing methods used in the investigations are divided into three categories: correlation analysis, rule-based methods, and machine-learning-based methods. The studies of correlation analysis found significant relationships between personal data characteristics, and indicated that a person’s personality traits can be used to predict their future job performance [19,20]. Recently, some researchers have considered the issue from a different perspective. Instead of determining what attributes are best for predicting personnel performance, they analyzed the performance gained by different classifiers. The rule-based methods use manually predefined rules which extract the useful patterns or rules to assist in personnel selection decisions. The limitation is that the classification performance will depend heavily on the ability of domain experts or professionals to design large bodies of rules.
In contrast to rule-based methods, machine-learning-based methods show better generalization. For instance, Cho et al. [21] predicted job performance of insurance sales agents using discriminant analysis and a decision tree. The method obtained a correct classification rate close to 70%, which could help insurance managers tp select quality agents in an insurance industry. Delgado-Gómez et al. [22] developed an expert system using support vector machines to forecast the sale performance of an insurance company. The system reached 5% higher accuracy than the state-of-the-art systems, which could lead to enormously reducing direct or indirect capital expenditure of the companies. Valle et al. [23] forecasted the job performance of sales agents in a call center using a naive Bayesian classifier. Each sample has an attribute vector that characterizes a sale agent. Attributes of each sample are as the input vector of the naive Bayesian model. Through supervised learning, the classifier predicts if the sale agent achieves a minimum production threshold. The classification problem is to judge whether the individual belongs to one class or another. This is a binary problem, and the model classified the instances with an accuracy of 80.60%. Thakur et al. [24] proposed a human performance prediction framework via ensemble-learning technology, which could help decision-makers to recruit a suitable employee in the software industry. Sarker et al. [25] used a hybrid method based on K-Means clustering and a decision tree to forecast staff performance, which could help executives to improve the quality of an organization. However, the biggest challenges of machine-learning-based methods are to choose good features which result in effective learning. Although many traditional classification algorithms have been presented to forecast the performance, the work mentioned above is still in its infancy, and there is great potential for achieving significant performance improvement.

2.2. Deep Learning

With the improvement in computing power and the availability of large datasets, deep learning technology has developed rapidly and has achieved unprecedented success in many other tasks. So far, no one has tried to apply deep learning to the personnel performance prediction task of HRM. However, the application of deep learning methods in the literature review is a great point of reference for us.
Deep learning is a family of state-of-the-art techniques. CNN is a popular deep learning technique. Krizhevsky et al. [26] trained a large-scale and deep convolutional neural network to classify images, which achieved the best performance compared to the previous work. CNN has also been used successfully for natural language processing (NLP), involving text classification and sentiment analysis [27,28]. Zhao et al. [29] employed a syntax word embedding to learn the syntactic information of a sentence and fed it into the CNN model to extract drug–drug interactions. This method achieved an F-score of 68.6%, which achieved a better performance than the previous state-of-the-art methods. Ombabi et al. [30] applied deep learning to the Twitter user’s comment classification, employed CNN to extract deep semantic features from the comments, and a support vector machine classifier was used to forecast the final classification. Finally, the topic in which users are interested was obtained, and the proposed method achieved a best accuracy of 97.3% for users’ interest classification. Studies have shown that CNNs have better results than traditional machine learning methods. Although CNNs have been shown to perform well, the classification model which only uses CNNs cannot learn the contextual dependency and structural information.
In addition to CNN models, recurrent neural network (RNN) models have achieved great success in many applications, e.g., object detection [31], video captioning [32], and drug–drug interaction extraction. Huang et al. [33] proposed a two-stage method for improving the performance of drug–drug interaction extraction, which obtained the highest F-score of 69.0%. In the first stage, it identified the positive instances employing a feature-based binary classifier. In the second stage, the LSTM-based classifier was used to classify the positive instances into a specific category. Recently, some researchers have also already successfully applied RNNs to some educational applications, such as student performance prediction, and have shown that they have a better performance than traditional machine learning methods. Deep knowledge tracing (DKT) used recurrent neural networks (e.g., RNN and LSTM) to model a student exercising process to forecast their performances [34]. Furthermore, by exploiting both student exercising records and the text descriptions of exercises, Su et al. [35] proposed an exercise-enhanced recurrent neural network (EERNN) framework for student performance prediction. A bidirectional LSTM was first designed to extract exercise semantic representations from texts. Then, a proposed LSTM architecture was used to trace student states for making final prediction.
Apart from the methods described above, some other works used the combination of CNN and RNN to do classification task. Other classification tasks can also be learned from. Wei et al. [36] employed a network based on CNN and LSTM to automatically identify the urgency of a post and classify the sentimental polarity. This approach used CNN to extract the local contextual feature representations of a post. Then, the post representations were fed into LSTM model to get the long temporal semantic information. The combination of CNN and RNN has also been successfully applied in many other computer tasks, e.g., direction of slip detection [37], multimodal wearable activity recognition [38], and human activity recognition in sensor-rich environments [39]. However, the classification model which only uses the hybrid model based on CNN and LSTM lacks the effective fusion of semantic and structural information. Further, attention-based methods are also popular. Guo et al. [40] used RNN to learn higher-level contextual representations and utilized CNN to gain sentence features for the relation classification task. The hybrid model integrated word-level and sentence-level attention mechanisms, which strengthened critical words and features and improved the performance.
Although deep learning methods have been widely applied in many tasks, for all we know, this is the first work to attempt personnel performance modeling based on deep learning. However, this personnel performance modeling research is still at an early stage, and its prediction performance has much room to improve.

3. Methods

In this paper, the prediction problem is treated as a classification problem, where a record with a prediction value 1 (0) indicates a positive (negative) instance. Public datasets and a personnel performance prediction dataset were used to train and test the proposed framework. The overview of our framework is illustrated in Figure 1. We propose a CHRNNA framework to achieve our goal. It contains the following steps.
(1)
Two kinds of neural networks, CNNs and RNNs, where the latter refers to LSTM, are employed in personnel prediction.This hybrid model combining CNN with LSTM is called the CRNN model;
(2)
The self-attention mechanism is introduced to a hybrid CRNN model, which aims to automatically capture the informative features;
(3)
The learned features, extracted from the last layer of CRNN model based on self-attention, are directly fed into the KNN classifier as inputs.
Next, we thoroughly describe the data and the two stages contained in the hybrid CRNN model with a self-attention mechanism and classification.

3.1. Personnel Performance Prediction Data

3.1.1. Data Description

To further estimate the performance of our method, we used a collected real dataset for a high-technology industry to forecast personnel performance. The characteristics of this dataset are shown in Table 1.
Next, the data collection and data preprocessing process are described in detail.

3.1.2. Data Collection

To be able to employ deep learning for modeling personnel performance, we required a dataset with labeled performance. Because there is no such human-labeled dataset publicly available, we should collect a personnel performance dataset with a performance-carrying label. The experimental dataset is collected from the Human Resource Department of a high-technology industry.
The first issue to consider is what attributes should be collected in this dataset. Salgado proposed the famous Big Five Model, which included five personality factors, i.e., conscientiousness, emotional stability, extraversion, openness, and agreeableness [41]. Güngör et al. thought that determining the most eligible person was dependent on some factors, such as work experience, foreign language, basic computer skill, personal goal, long life learning, etc. [42]. Li et al. pointed out that human and task characteristics are related to a person’s performance. We adopted an expert panel discussion and behavioral event interviews as the main method of collecting data. The panel of experts conducted a content analysis of the interview content to determine the competency characteristics exhibited by the respondents. Many aspects like “self-control”, “confidence”, “initiative”, “self-motivation”, and so on were also identified. After reaching an agreement through discussion, each evaluator was scored. That is, the determination of the attributes values was carried out by the expert group to analyze the content of the interviews. In order to make the collected data satisfy our demands, we created a personnel performance dataset including 5 categories and 22 attributes. These attributes were deemed to have an impact on personnel performance and are shown in detail in Table 1.
The second issue to contemplate is how to describe a person’s performance in this dataset. Just like many decision problems, the personnel performance problem is too sophisticated in real life. Since human behaviors and characteristics are complicated, it is difficult to quantify a person’s performance. People usually forecast inaccurately for quantitative problems, while relatively having an accurate prediction for qualitative problems. Therefore, we used qualitative fuzzy levels to describe attributes.
The components of the dataset are shown in Table 2. It contains 23 items, which includes 22 attributes and 1 personnel performance class. Each feature is represented by 5 fuzzy levels. For instance, a person’s memory capability has been divided into five fuzzy levels, with “very poor”, “poor”, “middle”, “good”, and “very good”. A person’s confidence has five fuzzy levels, which are “very unconfident”, “unconfident”, “medium”, “confident”, and “very confident”. Further, we used five fuzzy levels to describe experience, representing “completely inexperienced”, “inexperienced”, “middle”, “experienced”, and “well experienced”.
For different types of tasks, various means can be used to evaluate the personnel’s performance. Taking a high-technology industry as an example, the performance of personnel is determined by the completion of tasks and the meetings. To facilitate performance prediction by employing a classification algorithm, actual performance values correspond and are transformed to 2 grades, and we used the 2 integers of 0 and 1 to represent different grades of performance in the sample data. That is to say, real performance values are 0 and 1. If personnel performance is achieved, performance value equals to 1, otherwise it equals to 0.

3.1.3. Data Preprocessing

The raw data included some samples that were not applicable. Since these samples would reduce the ability to build a model, we needed to clean the data and remove all the duplicate data. Then, we manually carried out a random inspection of 300 instances from the dataset and found no duplicates. The initial dataset has 1151 samples. After discarding anomalous samples from the dataset, the number of applicable instances was reduced to 1139. The anomalous samples refer to duplicates and the data whose attributes value are all five. In this way, the dataset we attained can utilize deep learning. Thus, we now turn to describing the deep learning method we adopted.

3.2. First Stage: A Hybrid CRNN Model with Self-Attention

We call the hybrid model of CNN combined with LSTM as a CRNN model. Next, we describe the details of this hybrid CRNN model with a self-attention mechanism.

3.2.1. CNN for Feature Extraction

A critical problem for personnel performance prediction is feature representation, whereas traditional feature selection methods rely mainly on human-designed features. On one hand, hand-crafted feature extraction would be too time-consuming. On the other hand, hand-crafted feature extraction depends on human experience and requires designers to have strong professional background knowledge. Thus, the classification performance will be affected. Obviously, the disadvantages of this approach are apparent.
Recently, the rapid development of deep neural networks has brought new inspiration to feature extraction. In this section, we used CNNs to model, which we shall now introduce.
Our network uses two convolutional layers and a max-pooling layer, as shown in Figure 1. The convolution layer is used to automatically capture features, and the max-pooling layer is utilized to automatically extract which features play key roles in personnel performance prediction. The first procedure in our model is to train the network by inputting the sample itself and its label. Convolution layers contain a series of feature maps, which are formed by sliding diverse kernels over an input sample. A max-pooling manipulation is employed to capture the most important feature by extracting the biggest value from a feature map. More details about models are described in Section 4.

3.2.2. LSTM for Contextual Information

When people perform tasks, they need some basic qualities, such as perception, learning, creativity, memory, engagement, experience, health, etc. These information sequences are rich in content, and the information has a complex temporal correlation with each other. For example, for creativity, perception, learning, and confidence used in the design process, we need to tackle multidimensional input information at the same time, because it is constantly changing. Moreover, people do not start to carry out their tasks from scratch. As humans do computational design, they perform it based on the previous information. People do not abandon all known information and start to perform tasks from scratch again. However, CNN is unsuitable for modeling temporal information or dependency. LSTM has a memory function, which is extraordinary suitable for addressing this problem.
In this section, for our modeling, we used long short-term memory (LSTM) [43], a variation of RNNs. For a mathematical notation, we denote scalars with an italic lower case (e.g., h), vectors with a bold lower case (e.g., h), and matrices with a bold upper case (e.g., U).
LSTM is a kind of neural network architecture that is especially adapted to model sequential information. As illustrated in Figure 2, LSTM includes an input gate, a forget gate, and an output gate. LSTM networks tackle the problem of long-term dependencies of features via enhancing a memory cell at each time step t. We used LSTM to model the contextual dependencies and semantic relevance from our datasets, as shown in Figure 1.
LSTM takes an input vector x t , a hidden vector h t 1 , and a memory cell state vector c t 1 and generates c t and h t through the following calculations:
i t = σ ( W i x t + U i h t 1 + b i )
f t = σ ( W f x t + U f h t 1 + b f )
o t = σ ( W o x t + U o h t 1 + b o )
n t = tanh ( W n x t + U n h t 1 + b n )
c t = f t c t 1 + i t n t
h t = o t tanh ( c t )
where i t , f t , o t , and c t are the input gates, forget gates, output gates, and memory cell, respectively. The n t represents a new memory cell vector with candidates which could be added to the state. σ ( · ) is logistic sigmoid function, and tanh ( · ) is hyperbolic tangent function. ⊙ refers to the element-wise multiplication operation. The LSTM parameters W, U are weights, and b is bias, where W k , U k , and b k are for k { i , f , o , n } .

3.2.3. Self-Attention

By combining convolutional layers, max-pooling layer, and recurrent neural networks, our model adopts the strength of both convolutional neural models and recurrent neural models. Moreover, we want to capture long-range dependencies from personnel performance information. Recently, a self-attention mechanism has led to new ideas for solving these problems.
We think that 5 categories and 22 attributes have impacts on personnel performance in this paper. However, different features have various importance. Therefore, we require a strategy to discriminate the importance of the 22 attributes. To forecast personnel performance, the self-attention mechanism has the ability to identify the significance of diverse attributes rather than considering each attribute on average. Thus, we used the self-attention mechanism for the hybrid CRNN model.
In this section, we thoroughly describe self-attention. After obtaining the contextual features of the input sample, the self-attention mechanism [44] is used to learn the weight coefficient, which reflects the importance of each feature in the sample. Suppose that for each instance, we have N series of output from N LSTM cells, then the self-attention can be formulated as:
c i = j = 1 N α i j v i
We compute the weight coefficient α i j of each v i according to the following formulation:
α i j = e x p ( v i j ) k = 1 N e x p ( v i k )
v i j = F s c o r e ( v i , v j )
where c i is the representation of the sequence as the weighted sum of hidden representation, α i j is the normalized importance, v i j indicates the score about the degree of dependency between v i , and v j , Fscore is a function to compute the score about v i and v j . v i j calculated by function F-score is normalized by softmax via Equation (8). The output of this self-attention mechanism is a weighted sum.
Self-attention is seen as a separate layer, mixed with the CNN and LSTM model, which is able to more fully integrate their respective strengths.

3.2.4. Model Training

In recent years, the cross-entropy loss function has been widely used as a loss function in the model training of various tasks. Cross-entropy refers to the gap between the true probability distribution and the predicted probability distribution. In model training processes, we expect the probability that the model predicts the instance to be as similar as possible to the true probability. The formulation is defined as:
C E ( p , y ) = i = 1 N y i l o g ( P )
where i denotes the sample, N represents the total number of the samples, y is the one-hot vector corresponding to the true category of the sample, and P is the predicted probability. In this study, we used cross-entropy as a loss function. We expect the cross-entropy loss function to be minimized, because the lower the cross-entropy is, the closer the prediction distribution obtained by our model is to the true distribution.

3.3. Second Stage: Classification

The previous stage (neural network training) can be viewed as preprocessing because it can be executed independently before the classification stage.
At present, our society is besieged with large-scale data, and there is usually no prior knowledge of data distribution. The KNN classifier is very suitable for addressing these data. Further, KNN can get more complex decision information than the activation function of the last layer of our hybrid CRNN model. The idea of proposing our method is that these processes can complement each other perfectly and obtain the synergy of large datasets.
When it comes to classifying samples that need to be classified, our solution is described below:
(1)
The raw instances are propagated via our proposed network, and their feature vectors are extracted from the last layer of the hybrid CRNN model with self-attention;
(2)
The learned features mentioned above are fed into the KNN classifier as inputs;
(3)
The distances of the samples are computed, and the nearest training samples belonging to the test samples are selected;
(4)
The conventional KNN classification is carried out within these chosen data.
Next, we did extensive experiments on a personnel performance dataset to validate the goodness of our approach. More details about experiments are given in Section 4.

4. Experiments

In this section, our experiment and discussion are detailed. It includes the experimental setup, comparison algorithms, evaluation criterion, various experimental results to evaluate the performance of the presented scheme, and analysis of the case study.

4.1. Experimental Setup

This subsection introduces our experimental settings in detail. All experiments were performed on Python programming language (version 3.5), employing TensorFlow (version 1.10) and Keras (version 2.2). The PC we used includes an Intel(R) Core(TM) i5-4570 CPU running at 3.20 GHz. For a higher computing performance, we used a NVIDIA GeForce GTX 750 GPU with the cuDNN library.
In this experiment, the dataset was divided into the training set (60%), test set (20%), and validation set (20%). The learning of the CHRNN (without a self-attention mechanism) and CHRNNA framework (with a self-attention mechanism) was carried out for 200 epochs and a mini-batch size of 128 samples. Layer 1 with a convolution operator has 13 filters of size 1 × 2; Layer 2 with a convolution operator contains 26 filters of size 1 × 2, followed by a max-pooling layer. The max-pooling layer has a kernel size 2 and stride step 1. The dimension of the LSTM layer is 64. The dropout rate was set to 0.5 to prevent overfitting. For the binary classification problem, we adopted the sigmoid activation function. We employed binary cross-entropy loss on sigmoid function and the Adam optimization method, which works better to avoid the gradient vanishing/exploding issues. In the final classification stage, we studied different values of the number of neighbors k. More precisely, the k value considered is 7. Table 3 lists our hyper-parameter setting.

4.2. Comparison Algorithms

To demonstrate that our model is an excellent model, we introduced several comparison algorithms. We selected some traditional models, such as C4.5, RIPPER (repeated incremental pruning to produce error reduction), CBA (classification based on associations), CMAR (classification based on multiple association rules), MKNN (modified KNN) [45] and EEKNN (entropy Euclidean distance KNN). In addition, we also selected some deep learning models (e.g., CNN, RNN, and LSTM) as comparison methods of personnel performance prediction. We considered these traditional models and deep learning models as baselines.
  • C4.5 is a popular and powerful decision tree classifier;
  • RIPPER is a traditional rule-based classifier, whose accuracy may not be as high in most cases;
  • CBA is an important association rule-based classifier, which generates all the association rules with certain support and confidence thresholds;
  • CMAR is an association rule-based classifier, which uses multiple rules for prediction;
  • MKNN is an approach based on local learning which first introduces a new similarity function;
  • EEKNN is a performance modeling method based on data mining. A proposed improved KNN algorithm was used to handle the personnel performance prediction problem;
  • CNN is a popular model which is used for feature learning;
  • RNN is a deep learning model for sequential data modeling. Here, we considered it as a comparison method;
  • LSTM is a variant of RNN. It is an advanced RNN architecture which prevents vanishing gradients or the phenomena of exploding;
  • CHRNN is our method. It is a general framework without a self-attention mechanism;
  • CHRNNA is our method. It is our final selected framework which integrates the self-attention mechanism.

4.3. Evaluation Criterion

To quantitatively evaluate the classification performance of the proposed framework, we took the accuracy (acc) as a performance measure. We used the acc to compare the classification performance of different algorithms. The index acc is defined by true positive (TP), false negative (FN), false positive (FP), and true negative (TN), which is computed based on Equation (11):
acc = TP + FN TP + FN + FP + TN
The index acc is in the range of [0,1]. The bigger the value of acc is, the better the performance of the classifier.
Moreover, we also used precision (P), recall (R), and the F1 value as evaluation metrics to evaluate the performance of our proposed methods in this paper. The formula for computing P, R, and F1 is as follows:
P = TP TP + FP
R = TP TP + FN
F 1 = 2 × P × R P + R

4.4. Experimental Results

4.4.1. Experimental Results on a Personnel Performance Prediction Dataset

To further demonstrate the effectiveness and applicability of the proposed method, we introduced several methods for comparison on a real personnel performance prediction dataset. We repeated our experiments 10 times and reported the average results using the acc metric, as shown in Figure 3. We collected data from a high-technology industry. All samples were divided into two categories—one achieved personnel performance, the other did not achieve personnel performance. The forecasted values were compared with the true values to validate the applicability of our framework.
The previous work is based on traditional machine learning methods, while our approaches used deep learning. This paper compares the methods of C4.5, RIPPER, CBA, CMAR, MKNN, EEKNN, CNN, RNN, and LSTM. Figure 3 shows the performance comparisons. Our methods of CHRNN and CHRNNA achieved 95.13% and 96.49%, respectively. For the traditional machine learning method, EEKNN achieved an accuracy of 86.84%. In comparison with the baseline scheme EEKNN, CHRNN generated up to 8.29% accuracy improvement. Moreover, thanks to the introduction of the self-attention strategy, our CHRNNA method generated up to 1.36% accuracy increment again. For deep learning methods, CNN and LSTM achieved an accuracy of 92.10% and 91.81%. Compared to CNN and LSTM, our CHRNNA increased by 4.39% and 4.68%. The best performance was achieved by CHRNNA. When we compare deep learning methods (e.g., CHRNN and CHRNNA) to the traditional method (e.g., EEKNN) and existing deep learning methods (e.g., CNN and LSTM), the experimental results show that our deep learning methods outperform other methods for personnel performance prediction.
As Table 4 shows, we also adopted P, R, and F1 score to assess the performance of our method. The P, R, and F1 of CHRNNA on personnel performance data were 97.07%, 96.14%, and 96.60%, respectively, which are all higher than existing methods. The higher metrics values of the CHRNNA framework show CHRNNA has a better performance. It shows that the CHRNNA framework is more reliable when applied to personnel performance prediction.
From Figure 3 and Table 4, we conclude that the CHRNNA framework generally performs better than other methods. The reasons behind these phenomena are as follows:
Firstly, for C4.5, it only produces a small set of rules which may result in many interesting and useful rules not being discovered. For CBA and CMAR, they generate a very large number of potential classification rules. These approaches do not easily identify the most effective rule for classifying a new case. They simply select a rule with a user-defined measure, such as confidence. Such a simple pick may affect the classification performance, and their confidence-based rule evaluation measure may lead to overfitting. The results of machine-learning-based methods are also not encouraging. However, our method is different. Our CRNN model can automatically learn discriminative features, which avoids inefficient and cumbersome process. The CRNN model applies a recurrent structure to better learn contextual information of features. When humans carry out tasks, they do not start from scratch. They need to deal with multiple input information of a high dimension at the same time, which has temporal correlation. CRNN has a memory function that is suitable for solving the problems of temporal information or dependence in personnel performance prediction. Our model uses the advantage of both convolutional neural models and recurrent neural models.
Secondly, KNN can attain more complex decision information than the activation function in the last layer of the hybrid CRNN model. This fact clearly shows the advantage of the presented combined method of the CRNN model and KNN algorithm compared to conventional machine learning methods. It is interesting that CRNN and KNN reinforce one another, which achieves a better classification performance.
Finally, we summarize that CHRNNA is superior over the general CHRNN framework. The application of our approach depends on the size and attribute diversity of the instances. Different features have different effects on personnel performance. We used self-attention strategy to discriminate the importance of various features. The most important features are the most prioritized information in the self-attention layer. We believe that the self-attention layer has the ability to learn long dependencies without considering the distance in sequences. The self-attention mechanism is helpful to achieve a high-quality prediction in accuracy.
In summary, we are content to find that deep learning techniques can help in improving the performance. To be precise, the self-attention mechanism can further improve the classification performance. We believe that our proposed method can be successfully applied to personnel performance prediction.

4.4.2. The Time Efficiency Analysis

Computation time is another way to compare different methods. The running time of these algorithms is shown in Table 5.
The running time of the EEKNN algorithm took 52.04 s, while CHRNNA was almost six times faster, 8.89 s. The total running time of the MKNN algorithm is the lowest. The running time in seconds shows that the CHRNNA method is significantly faster than the MKNN and EEKNN algorithms. Although our proposed method is somewhat slower than several methods in running time, it has a significant performance improvement in terms of acc, P, R, and F1 evaluation indices. With the continuous improvement of computing power, it can make up for the lack of response time. Thus, the difference in running time is negligible. Combining comparison results in terms of acc, P, R, F1, and running time, we can conclude that the CHRNNA method is more effective than the compared algorithms.

4.5. Ablation Study of Various Modules

To further evaluate the effectiveness of various modules, we reported the ablation study of our CHRNNA model. Our main model outperforms the other variants. We applied our CHRNNA model and its two submodels, which are without self-attention and LSTM. The experimental results for the ablation study are listed in Table 6. The results of the “- self-attention” row refer to the results without the self-attention mechanism. The results of the “- LSTM” row refer to the results without LSTM. The results indicate that: (1) our main model is better than the two variants under the same settings; and (2) as expected, the results for the simplified models all drop a lot. This clearly demonstrates the effectiveness of these modules. For the submodel “- self-attention”, the drop of ACC, P, R, and F1 demonstrates that the self-attention mechanism is critical. Specifically, the self-attention mechanism helps to evaluate the contribution of various attributes, so we can identify the most influential factor for the final performance decision. The drop in the performance of “- LSTM” also shows that LSTM is important for achieving a good performance.

5. Conclusions

High-technology industries depend on personnel performance to maintin their competitive advantages. In this paper, we presented a novel CHRNNA framework, which is used to forecast personnel performance in the future and help decision-makers to select the most adequate talents. We designed and collected a dataset with 22 attribute items, which were used to reflect personnel performance.
The proposed framework remarkably improves prediction performance. In the first stage, we employ a hybrid CRNN model with a self-attention mechanism to automatically capture the global informative features. In the second stage, we use a KNN-based classifier to forecast personnel future performance. Experimental results demonstrate that our proposed method yields a significant performance.
In our opinion, there is some room for further improvement with minimal accuracy losses, such as studying more advanced network architectures and designing a loss function, which will be the focus of our future work.

Author Contributions

Methodology, performing experiments, and writing—original manuscript, X.X.; Conceptualization and writing—review and editing, J.F.; Debugging code and validation, Y.G.; Performing experiments, M.L.; Writing—editing, W.Z. and X.S.; Data preprocessing and validation, A.Z. and S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by the Key Research and Development Program in the ShaanXi Province of China (No. 2019ZDLGY03-10), the National Natural Science Foundation Projects of China (No. 61877050), the Major Issues of Basic Education in the Shaanxi Province of China (No. ZDKT1916) and the Scientific research project of Education Department of the Shaanxi Provincial Government (No. 19JK0808).

Acknowledgments

The authors would like to thank the anonymous reviewers for the constructive comments. These comments are valuable and helpful for revising and improving our paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bal, M.; Bal, Y.; Ustundag, A. Knowledge Representation and Discovery Using Formal Concept Analysis: An HRM Application. World Congress Eng. 2011, 2, 1068–1073. [Google Scholar]
  2. Chien, C.-F.; Chen, L.-F. Data mining to improve personnel selection and enhance human capital: A case study in high-technology industry. Expert Syst. Appl. 2008, 2, 280–290. [Google Scholar] [CrossRef]
  3. Karahoca, A.; Karahoca, D.; Kaya, O. Data mining to cluster human performance by using online self regulating clustering method. In Proceedings of the Wseas International Conference on Multivariate Analysis & Its Application in Science & Engineering, Istanbul, Turkey, 27–30 May 2008. [Google Scholar]
  4. Gobert, J.D.; Sao Pedro, M.A.; Baker, R.S.J.D.; Toto, E.; Montalvo, O. Leveraging Educational Data Mining for Real-Time Performance Assessment of Scientific Inquiry Skills within Microworlds. J. Educ. Data Mining 2012, 4, 104–143. [Google Scholar]
  5. Li, N.; Kong, H.; Ma, Y.; Gong, G.; Huai, W. Human performance modeling for manufacturing based on an improved KNN algorithm. Int. J. Adv. Manuf. Technol. 2016, 84, 473–483. [Google Scholar] [CrossRef]
  6. Wang, Q.; Li, B.; Hu, J. Feature Selection for Human Resource Selection Based on Affinity Propagation and SVM Sensitivity Analysis. In Proceedings of the World Congress on Nature & Biologically Inspired Computing, Coimbatore, India, 9–11 December 2009. [Google Scholar]
  7. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  8. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  9. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  10. Wang, H.; Feng, J.; Zhang, Z.; Su, H.; Cui, L.; He, H.; Liu, L. Breast mass classification via deeply integrating the contextual information from multi-view data. Pattern Recognit. 2018, 80, 42–52. [Google Scholar] [CrossRef]
  11. Xu, X.; Li, G.; Xie, G.; Ren, J.; Xie, X. Weakly Supervised Deep Semantic Segmentation Using CNN and ELM with Semantic Candidate Regions. Complexity 2019, 2019, 9180391. [Google Scholar] [CrossRef]
  12. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  14. Satt, A.; Rozenberg, S.; Hoory, R. Efficient Emotion Recognition from Speech Using Deep Learning on Spectrograms. In Proceedings of the 18th Annual Conference of the International Speech Communication Association (Interspeech 2017), Stockholm, Sweden, 20–24 August 2017; pp. 1089–1093. [Google Scholar]
  15. Bartz, C.; Herold, T.; Yang, H.; Meinel, C. Language Identification Using Deep Convolutional Recurrent Neural Networks. In Proceedings of the 24th International Conference on Neural Information Processing (ICONIP 2017), Guangzhou, China, 14–18 November 2017; pp. 880–889. [Google Scholar]
  16. Zhang, J.; Zhu, Y.; Zhang, X.; Ye, M.; Yang, J. Developing a Long Short-Term Memory (LSTM) based model for predicting water table depth in agricultural areas. J. Hydrol. 2018, 561, 918–929. [Google Scholar] [CrossRef]
  17. Sun, X. Drug-Drug Interaction Extraction via Recurrent Hybrid Convolutional Neural Networks with an Improved Focal Loss. Entropy 2019, 21, 37. [Google Scholar] [CrossRef] [Green Version]
  18. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  19. Barrick, M.R.; Mount, M.K. The big five personality dimensions and job performance: A meta-analysis. Personnel Psychology 1991, 44, 1–26. [Google Scholar] [CrossRef]
  20. Borman, W.C.; Motowidlo, S.J. Task Performance and Contextual Performance: The Meaning for Personnel Selection Research. Hum. Perform. 1997, 10, 99–109. [Google Scholar] [CrossRef]
  21. Cho, V.; Ngai, E.W.T. Data mining for selection of insurance sales agents. Expert Syst. 2003, 20, 123–132. [Google Scholar] [CrossRef]
  22. Delgado-Gómez, D.; Aguado, D.; Lopez-Castroman, J.; Cruz, C.S.; Artés-Rodríguez, A. Improving sale performance prediction using support vector machines. Expert Syst. Appl. 2011, 38, 5129–5132. [Google Scholar] [CrossRef]
  23. Valle, M.A.; Varas, S.; Ruz, G.A. Job performance prediction in a call center using a naive Bayes classifier. Expert Syst. Appl. 2012, 39, 9939–9945. [Google Scholar] [CrossRef]
  24. Thakur, G.S.; Gupta, A.; Gupta, S. Data Mining for Prediction of Human Performance Capability in the Software Industry. Int. J. Data Mining Knowl. Manag. Process 2015, 5, 53–64. [Google Scholar] [CrossRef] [Green Version]
  25. Sarker, A.; Shamim, S.M.; Zama, P.D.M.S.; Rahman, M.M. Employee’s Performance Analysis and Prediction Using K-means Clustering & Decision Tree Algorithm. Glob. J. Comput. Sci. Technol. 2018, 18, 1–6. [Google Scholar]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  27. Kalchbrenner, N.; Grefenstette, E.; Blunsom, P. A Convolutional Neural Network for Modelling Sentences. In Proceedings of the Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014, Baltimore, MD, USA, 22–27 June 2014. [Google Scholar]
  28. Ren, Y.; Zhang, Y.; Zhang, M.; Ji, D. Context-Sensitive Twitter Sentiment Classification Using Neural Network. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  29. Zhao, Z.; Yang, Z.; Luo, L.; Lin, H.; Wang, J. Drug drug interaction extraction from biomedical literature using syntax convolutional neural network. Bioinformatics 2016, 32, 3444–3453. [Google Scholar] [CrossRef]
  30. Ombabi, A.H.; Lazzez, O.; Ouarda, W.; Alimi, A.M. Deep learning framework based on Word2Vec and CNN for users interests classification. In Proceedings of the 2017 Sudan Conference on Computer Science and Information Technology (SCCSIT), Elnihood, Sudan, 17–19 November 2017; pp. 1–7. [Google Scholar]
  31. Li, X.; Ye, M.; Liu, Y.; Zhang, F.; Liu, D.; Tang, S. Accurate object detection using memory-based models in surveillance scenes. Pattern Recognit. 2017, 67, 73–84. [Google Scholar] [CrossRef]
  32. Yang, Y.; Zhou, J.; Ai, J.; Bin, Y.; Hanjalic, A.; Shen, H.T.; Ji, Y. Video Captioning by Adversarial LSTM. IEEE Trans. Image Process. 2018, 27, 5600–5611. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Huang, D.; Jiang, Z.; Zou, L.; Li, L. Drug–drug interaction extraction from biomedical literature using support vector machine and long short term memory networks. Inf. Sci. 2017, 415–416, 100–109. [Google Scholar] [CrossRef]
  34. Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.; Guibas, L.; Sohl-Dickstein, J. Deep knowledge tracing. In Proceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 505–513. [Google Scholar]
  35. Su, Y.; Liu, Q.; Liu, Q.; Huang, Z.; Yin, Y.; Chen, E.; Ding, C.H.Q.; Wei, S.; Hu, G. Exercise-Enhanced Sequential Modeling for Student Performance Prediction. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), New Orleans, LA, USA, 2–7 February 2018; pp. 2435–2443. [Google Scholar]
  36. Wei, X.; Lin, H.; Yang, L.; Yu, Y. A Convolution-LSTM-Based Deep Neural Network for Cross-Domain MOOC Forum Post Classification. Information 2017, 8, 92. [Google Scholar] [CrossRef] [Green Version]
  37. Zapata-Impata, B.S.; Gil, P.; Torres, F. Learning Spatio Temporal Tactile Features with a ConvLSTM for the Direction Of Slip Detection. Sensors 2019, 19, 523. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Ordóñez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Baldominos, A.; Saez, Y.; Isasi, P. Evolutionary Design of Convolutional Neural Networks for Human Activity Recognition in Sensor-Rich Environments. Sensors 2018, 18, 1288. [Google Scholar] [CrossRef] [Green Version]
  40. Guo, X.; Zhang, H.; Yang, H.; Xu, L.; Ye, Z. A Single Attention-Based Combination of CNN and RNN for Relation Classification. IEEE Access 2019, 7, 12467–12475. [Google Scholar] [CrossRef]
  41. Salgado, J.F. The five factor model of personality and job performance in the European Community. J. Appl. Psychol. 1997, 82, 30–43. [Google Scholar] [CrossRef]
  42. Güngör, Z.; Serhadlioglu, G.; Kesen, S.E. A fuzzy AHP approach to personnel selection problem. Appl. Soft Comput. 2009, 9, 641–646. [Google Scholar] [CrossRef]
  43. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  44. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  45. Bhattacharya, G.; Ghosh, K.; Chowdhury, A.S. An affinity-based new local distance function and similarity measure for kNN algorithm. Pattern Recognit. Lett. 2012, 33, 356–363. [Google Scholar] [CrossRef]
Figure 1. The architecture of our framework.
Figure 1. The architecture of our framework.
Entropy 21 01227 g001
Figure 2. Long short-term memory (LSTM) structure.
Figure 2. Long short-term memory (LSTM) structure.
Entropy 21 01227 g002
Figure 3. Performance of all the methods on personnel performance prediction.
Figure 3. Performance of all the methods on personnel performance prediction.
Entropy 21 01227 g003
Table 1. The personnel performance dataset used for experimentation.
Table 1. The personnel performance dataset used for experimentation.
DatasetsInstancesAttributesClasses
Personnel Performance1139222
Table 2. The components of the personnel performance dataset.
Table 2. The components of the personnel performance dataset.
CategoryAttributesValueDescription
CapabilityPerception1–55 levels of perception
 Learning1–55 levels of learning
Memory1–55 levels of memory
Creativity1–55 levels of creativity
Speech1–55 levels of speech
Logic1–55 levels of logic
ConscientiousnessDedication1–55 levels of dedication
Engagement1–55 levels of engagement
Self control1–55 levels of self control
Decision making1–55 levels of decision making
AchievementGoal1–55 levels of goal
Initiative1–55 levels of initiative
Independence1–55 levels of independence
Confidence1–55 levels of confidence
EmotionEmotion awareness1–55 levels of emotion awareness
Emotion expression1–55 levels of emotion expression
Self-motivation1–55 levels of self-motivation
Emotion management1–55 levels of emotion management
Complementary qualityEducation1–55 levels of education
Health1–55 levels of health
Experience1–55 levels of experience
Stress tolerance1–55 levels of stress tolerance
Personnel performance 0, 10, not achieve performance; 1, achieve performance
Table 3. Parameter settings.
Table 3. Parameter settings.
ParametersParameters Value
epochs200
batch size128
convolution 1 layer: filters number13
convolution 2 layer: filters number26
filters size1 × 2
LSTM dimension64
Dropout rate0.5
Table 4. Performance of comparison methods on personnel performance data in terms of precision (P), recall (R), and F1 (%). The best experimental results are given in bold.
Table 4. Performance of comparison methods on personnel performance data in terms of precision (P), recall (R), and F1 (%). The best experimental results are given in bold.
MetricsC4.5RIPPERCBACMARMKNNEEKNNCNNRNNLSTMOurs
CHRNNCHRNNA
P71.4782.5280.8657.1667.2271.9392.1489.4192.0696.0297.07
R78.9285.0188.3679.2860.5467.1292.5489.0491.6794.6196.14
F174.3283.7084.1160.0262.6869.1292.3489.2291.8695.3196.60
Table 5. Computation time of contrastive experiments measured in seconds per run.
Table 5. Computation time of contrastive experiments measured in seconds per run.
C4.5RIPPERCBACMARMKNNEEKNNCNNRNNLSTMCHRNNCHRNNA
0.825.444.5211.311878.5652.045.073.276.5612.298.89
Table 6. Ablation study of CHRNNA with various modules on personnel performance. The best results are given in bold.
Table 6. Ablation study of CHRNNA with various modules on personnel performance. The best results are given in bold.
StrategyACCΔ1PΔ2RΔ3F1Δ4
CHRNNA (Our Model)96.49-97.07-96.14-96.60-
- Self-attention95.13−1.3696.02−1.0594.61−1.5395.31−1.29
- LSTM95.83−0.6696.04−1.0395.61−0.5395.81−0.79

Share and Cite

MDPI and ACS Style

Xue, X.; Feng, J.; Gao, Y.; Liu, M.; Zhang, W.; Sun, X.; Zhao, A.; Guo, S. Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction. Entropy 2019, 21, 1227. https://doi.org/10.3390/e21121227

AMA Style

Xue X, Feng J, Gao Y, Liu M, Zhang W, Sun X, Zhao A, Guo S. Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction. Entropy. 2019; 21(12):1227. https://doi.org/10.3390/e21121227

Chicago/Turabian Style

Xue, Xia, Jun Feng, Yi Gao, Meng Liu, Wenyu Zhang, Xia Sun, Aiqi Zhao, and Shouxi Guo. 2019. "Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction" Entropy 21, no. 12: 1227. https://doi.org/10.3390/e21121227

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop