Next Article in Journal
Measuring Ethical Values with AI for Better Teamwork
Previous Article in Journal
ReSQoV: A Scalable Resource Allocation Model for QoS-Satisfied Cloud Services
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Clinical Trial Classification of SNS24 Calls with Neural Networks

1
Department of Computer Science, University of Évora, 7000-671 Évora, Portugal
2
Department of Computer Science, Zhongyuan University of Technology, Zhengzhou 450007, China
3
Centro ALGORITMI, Vista Lab, University of Évora, 7000-671 Évora, Portugal
4
CIDEHUS, University of Évora, 7000-809 Évora, Portugal
5
Serviços Partilhados do Ministério da Saúde, 1050-099 Lisboa, Portugal
*
Authors to whom correspondence should be addressed.
Future Internet 2022, 14(5), 130; https://doi.org/10.3390/fi14050130
Submission received: 6 April 2022 / Revised: 22 April 2022 / Accepted: 24 April 2022 / Published: 26 April 2022
(This article belongs to the Section Big Data and Augmented Intelligence)

Abstract

:
SNS24, the Portuguese National Health Contact Center, is a telephone and digital public service that provides clinical services. SNS24 plays an important role in the identification of users’ clinical situations according to their symptoms. Currently, there are a number of possible clinical algorithms defined, and selecting the appropriate clinical algorithm is very important in each telephone triage episode. Decreasing the duration of the phone calls and allowing a faster interaction between citizens and SNS24 service can further improve the performance of the telephone triage service. In this paper, we present a study using deep learning approaches to build classification models, aiming to support the nurses with the clinical algorithm’s choice. Three different deep learning architectures, namely convolutional neural network (CNN), recurrent neural network (RNN), and transformers-based approaches are applied across a total number of 269,654 call records belonging to 51 classes. The CNN, RNN, and transformers-based model each achieve an accuracy of 76.56%, 75.88%, and 78.15% over the test set in the preliminary experiments. Models using the transformers-based architecture are further fine-tuned, achieving an accuracy of 79.67% with Adam and 79.72% with SGD after learning rate fine-tuning; an accuracy of 79.96% with Adam and 79.76% with SGD after epochs fine-tuning; an accuracy of 80.57% with Adam after the batch size fine-tuning. Analysis of similar clinical symptoms is carried out using the fine-tuned neural network model. Comparisons are done over the labels predicted by the neural network model, the support vector machines model, and the original labels from SNS24. These results suggest that using deep learning is an effective and promising approach to aid the clinical triage of the SNS24 phone call services.

1. Introduction

SNS24, the Portuguese National Health Contact Center, is a national telephone and digital public service in Portugal. SNS24 allows citizens access facility, guarantees equity, and simplifies access to the National Health System (SNS). SNS24 provides clinical services for the citizens, such as triage, counseling, referral, and non-clinical help. These services are essential to the public health and safety. Health professionals and specifically nurses with distinct training supplied by SNS24 provide the telephone clinical services.
Currently, 59 possible clinical algorithms are developed by health professionals and approved by the Directorate General of Health (DGS). Following the pre-defined clinical pathways, the health professional or nurse selects the most appropriate one according to the citizen’s self-reported symptoms and relevant information about medical history. The selection of the clinical algorithms leads to five possible final referrals: self-care, clinical assessment at a primary health care center, clinical assessment in hospital emergency, transference to the National Medical Emergency Institute (INEM), or transference to the Poison Information Center. All the 59 clinical pathways start by asking screening questions, aiming to identify critical situations which need immediate health care and that are rapidly transferred to National Medical Emergency Institute. For non-emergencies, nurses must ensure that the main symptoms reported by the citizen are correctly mapped to allow the choice of the clinical pathway. Age, gender, and clinical history should also be considered during this process since this information is often crucial to the selection of the most appropriate clinical pathway. The choice of the clinical pathway by the health care professionals in each telephone triage episode is extremely important since it will determine the final referral. These pathways use a risk-averse system of prioritization, as other triage protocols such as the Manchester Triage System [1]. For this reason, SNS24 has an important role in the identification of users’ clinical situation and the correct referral to health services according to their symptoms.
This work explores using neural network (NN) architectures to build classification models for SNS24 clinical triage, aiming to improve the quality and performance of telephone triage service; support the nurses in the triage process, particularly with clinical algorithm’s choice; decrease the duration of the phone calls and allow a finer and faster interaction between citizens and SNS24 service. The main contributions of this paper are summarized as follows:
  • Three deep learning (DL) based classification architectures are compared, namely CNN (convolutional neural network), RNN (recurrent neural network), and transformers-based architecture, in the automatic clinical trial classification of the SNS24 health calls.
  • An empirical process for fine-tuning hyperparameters is explored. This process practically shortens the tuning time and theoretically achieves near-optimal classification accuracy.
  • Comprehensive experimental results show that neural network models outperform shallow machine learning models with the testing accuracies on the same SNS24 clinical dataset.
  • Useful suggestions are provided to the SNS24 health center. Practical analysis among similar clinical symptoms is carried out. Predictions from SNS24 triage, neural network models, and shallow machine learning models are compared.
The rest of the paper is organized as follows. Section 2 reviews the related work, Section 3 describes the materials used and presents the methods proposed. The experiments are demonstrated in Section 4, and the results and discussion are presented in Section 5. Section 6 analyzes similar clinical symptoms based on the built neural network models. Section 7 compares the labels predicted by the trained models with the original labels. Finally, the paper is concluded in Section 8.

2. Literature Review

Applying machine learning and deep learning technologies in healthcare domain is an active research topic [2]. A wide variety of paradigms, such as linear, probabilistic, and neural networks, has been proposed for clinical classification problems. For examples, Spiga et al. [3] used machine learning techniques for patient stratification and phenotype investigation in rare diseases; Sidey et al. [4] used machine learning techniques for cancer diagnosis. While applying traditional machine learning methods on clinical text classification, feature engineering is generally carried out using different forms of knowledge sources or rules. However, these approaches usually could not automatically learn effective features.
Recently, neural network methods which show capable feature learning ability have been successfully applied to clinical domain classification. Some works have also compared the performance between deep learning and shallow machine learning approaches and presented better performances when using the former approaches.
Research work by Shin et al. [5] shows that the overall performance of the systems can be improved using Deep Learning architectures. They used neural network models to classify radiology electronic health records. They compared two neural network classification models, CNN (convolution neural network) and NAM (neural attention mechanism), with the baseline SVM model. On average, the two neural models (CNN and NAM) achieved an accuracy of 87% and an improvement of 3% over the SVM baseline model.
Wu and Wang [6] used CNN in predicting the single underlying cause of death from a list of relevant medical conditions. They worked on a dataset containing 1,499,128 records and 1180 possible classes as causes of death. They compared the CNN models to several shallow classifiers (SVM, naïve bayes, random forest, and traditional BoW classification techniques). The CNN model achieved around 75% accuracy and was able to outperform all other shallow classifier models.
Mullenbach et al. [7] presented a convolutional neural network for multi-label document classification, aiming to predict the most relevant segments for the possible medical codes from clinical texts. They evaluated the approach on the MIMIC-II and MIMIC-III datasets, two open-access datasets of texts and structured records from the hospital ICU. Their method obtained a precision of 0.71 and a micro-F1 of 0.54 on the MIMIC-III dataset, and a precision of 0.52 and a micro-F1 of 0.44 on the MIMIC-II dataset.
Hughes et al. [8] used a CNN-based approach for sentence-level classification of medical documents into one of the 26 categories. Their results showed that when compared with the shallow learning methods, the CNN-based approach captured more complex features to represent the semantics of the sentence. The CNN-based model achieves an accuracy of 68%, the highest score among all the tested models.
Baker et al. [9] researched on using CNN models to classify biomedical texts, and the experiments were carried out on a cancer related dataset. Their evaluations showed that a basic CNN model achieved competitive performance compared with an SVM (Support Vector Machine) trained using manually optimized engineered features. The CNN-based model outperformed the SVM with modifications to the CNN hyperparameters, initialization, and training process.
Zhou et al. [10] experimented an integrated CNN-RNN model to provide patients with pre-diagnosis suggestions or clinic guidance online. Their experiments were carried out on the available online medical data. CNN, RNN, RCNN (Region CNN), CRNN (Convolutional RNN), and CNN-RNN models were compared. The integrated CNN-RNN model improved classification precision and was efficient in training efficiency.
Gao et al. [11] explored using CNN, RNN, and RCNN in clinical department recommendations. The experiments were carried out on a local constructed corpus consisting with 20 thousands patient symptom descriptions. The RCNN model showed better accuracy than when compared to CNN or RNN models, achieving an accuracy score of 76.51%.
In more recent research, transformers were compared against CNN models and hierarchical self-attention networks (HiSAN) [12]. The experimental results showed that generally, the CNN and HiSAN models achieved better performance than the BERT (Bidirectional Encoder Representations from Transformers) models.
Behera et al. [13] compared five deep learning based classification models, including DNN, RNN, CNN, RCNN, and RMDL (Random Multimodel Deep Learning), in automatic classification of four benchmark biomedical datasets. Among all deep learning models, the RMDL model provides the best classification performance on three datasets, and the RCNN classifier performs best on one dataset. Moreover, the work compared the performance between deep learning and shallow learning algorithms and showed that all the deep learning algorithms provided better classification performances.
Al-Garadi et al. [14] explored using text classification approaches for the automatic detection of non-medical prescription medication usage. Their work compared transformers-based language models, fusion-based approaches, several traditional machine learning and deep learning approaches. The results showed that the BERT and fusion-based models outperformed the others using machine learning and other deep learning techniques.
Mascio et al. [15] explored using several word representations and classification approaches for clinical text classification. They experimented and analyzed the impact on MIMIC-III and CLEF ShARe datasets. The results showed that the tailored traditional approaches of Word2Vec, FastText or GloVe were able to obtain or exceed the BERT contextual embeddings.
Flores et al. [16] compared a group of approaches, including an active learning approach, SVMs, Naïve Bayes, and a BERT classifier, on three datasets for biomedical text classification. The active learning approach obtained an AUC (areas under the curve) greater than 0.85 in all cases, being able to more efficiently reduce the number of training examples for equal performance than the other classifiers.
These related works above present an idea of applying neural network methodology in the area of clinical triage. As we can see, deep learning approaches such as CNN, RNN, and other models have been widely applied in the area of clinical text classification, and have shown powerful feature learning capability and play an important role in text classification. Exploring deep learning approaches for SNS24 clinical triage classification is a meaningful contribution, and having better-performed models will have an important impact on the SNS24 clinical triage service.

3. Materials and Methods

3.1. Dataset

SPMS (Serviços Partilhados do Ministério da Saúde) https://www.spms.min-saude.pt/ (accessed on 2 April 2022) cooperates, shares knowledge, and develops activities in the domains of health information and communication systems, making sure that all information is available in the best way for all citizens. SPMS advocates the definition and usage of standards, methodologies, and requirements. This guarantees the interoperability and interconnection among the health information systems, and as well with cross-sectional information systems of the Public Administration. The SPMS provided the anonymized data, and the competent ethics committee approved the study protocol used in this paper.
The original dataset used for our task was collected by the SNS24 phone-line from January to March 2018. It contains information of call records received, which has a total of 269,658 records with 18 fields. Table 1 lists all 18 fields both in the original Portuguese and in English. Each call record includes personal data, such as age, gender, and encrypted primary care unit. It also contains the calling information, such as the start/end time, initial intention, comments, contact reason, clinical pathway, and final disposition. The “contact reason” and “comments” fields are free text written in Portuguese by the technician or nurse who answered the respective call. The date related information is recorded in date format and the other remaining fields are nominal attributes.
Within the original 3-month dataset, there are 52 clinical pathways from the total of 59 defined by the SPMS. The proportions for each clinical pathway varies significantly. For example, Tosse (Cough) has the highest number of records and accounts for 14.006% of the calls, while Pr. por calor (Heat problems) represents only 0.001% of the calls; six of the clinical pathways have over 10,000 calls, and three under 100. Table A1, in the Appendix A, lists the existing 52 clinical pathways, their record numbers, and the proportions to the whole dataset.
While building the dataset to be used for this research, instances belonging to one clinical pathway with a number smaller than 50 were discarded. Thus, instances belonging to Pr. por calor were removed (see Table A1). As a result, the final dataset used for our task is composed of 269,654 records belonging to the 51 clinical pathways. Following conclusions from previous experiments [17], the “contact reason” field was selected as the discriminant information for our experiments; other fields, such as age, gender, etc. were not used. The “contact reason” field is one of the 18 descriptive attributes. It is written with a medium-length free text, consisting of straightforward information about the patient’s problem. Table A2 (in the Appendix A) presents examples of “contact reason” field texts from the five most frequent clinical pathways and their respective label. As a summary, the statistics of the original and experimental dataset are presented in Table 2.

3.2. Methods

Our main task is to support the nurses with the clinical algorithm’s choice. This can be framed as a supervised multi-class classification problem using as input the dataset we obtained from SNS24 calling records. The existing clinical dataset includes clinical texts and their corresponding clinical pathways are used for training a deep neural network model. Given a new record, the attribute “clinical pathway” is the class aiming to be predicted. As Figure 1 presents, the trained neural network model is used to predict the clinical pathway given the information collected by the nursers, and the predicted clinical pathway is finally returned to the nurse as the suggestion.
Figure 2 presents the general framework for training a neural network model in our task. The design of a neural network model is highly depending on the selected architecture, such as the number of layers, the design of hidden layers, the pre-trained language model selection, etc. We propose to train neural network models using CNN, RNN, and transformers-based architectures:
  • CNN-based architecture. We use the architecture described by Yoon Kim [18], which is a classical convolutional neural network for text classification.
  • RNN-based architecture. Mainly, two prevailing RNN types have been developed from the basic RNN [19]: long short-term memory (LSTM) [20] and gated recurrent unit (GRU) [21]. We use GRU type RNN in our experiments.
  • Transformers-based architecture. Transformers is a sequence-to-sequence architecture that was originally proposed for neural machine translation, and it shows effective performance in natural language processing tasks [22].
As Figure 2 shows, for each architecture, “contact reason” and “clinical pathway” are used as the articles and labels. They are pre-processed and fed to the neural network as the input. When embedding the input “contact reason” texts and building other hidden layers, pre-trained language models are used depending on the neural network architecture selection.
Compared to the one-hot representation which maps words into high-dimensional and sparse data, the static distributed representation (word embeddings) maps the words into a low-dimensional continuous space and can capture the semantic meanings of words. In the scope of static distributed representations, a number of pre-trained word embeddings are available, such as Word2Vec, Glove, fastText, etc. More recently, the contextualized (dynamic) distributed representation (language models) presents the ability to encode the semantic and syntactic information [23]. For example, pre-trained language models include ELMO [24], BERT [25], XLNet [26], Flair [27,28], etc. Since both the pre-trained word embeddings and pre-trained language models are state-of-the-art techniques to transform word into distributed representations, we take both into account in our task.

4. Experiments

In this section, we start by describing the experimental setup. As the first step, we preliminarily experiment with all three neural network architectures over a small number of epochs. Based on the results obtained from the preliminary experiments, the following step is to select the best-performed architecture among all three and further fine-tuning the models with selected hyperparameters step by step. The final best fine-tuned model will be used for class prediction.

4.1. Experimental Setup

In our previous work, we used shallow machine learning approaches to predict the clinical pathways on the SNS24 dataset [17]. Different from deep learning approaches, shallow machine learning approaches or conventional machine learning approaches usually need human intervention in feature extraction. Typical shallow machine learning approaches include Linear/Logistic Regression, Decision Tree, SVM, Naive Bayes, Random Forest, etc.
To be able to compare to our previously obtained results using shallow machine learning approaches [17], we use the same train/validation/test splits of the same dataset. The dataset is stratified split into the train (64%), validation (16%), and test set (20%). The neural network models are trained over the train set and adjusted over the validation set, and the test set is used for the final evaluation of the models.
Also, we use the same field texts as used in the shallow machine learning approach. The original dataset contains 18 fields (attributes), and only the texts of the “contact reason” field and the labels of the “Clinical pathway” field are used in building a prediction model. The texts from the other fields of the dataset are not used. Example texts and labels are shown in Table A2.
We use TensorFlow 2.2, Pytorch 1.8, and Python 3.8 as the framework in building the neural network models. We mainly use two state-of-the-art NLP libraries and their pre-trained language models: Transformers https://huggingface.co/transformers/ (accessed on 2 April 2022) and FLAIR https://github.com/zalandoresearch/flair (accessed on accessed on 2 April 2022).

4.2. Preliminary Experiments

4.2.1. CNN-Based Architecture

Among a number of the convolutional neural networks (CNN) that have presented excellent performances in many natural language processing tasks, one typical CNN architecture for text classification is proposed by Kim [18]. This type of CNN architecture generally includes these layers: embedding layer (input layer), convolutional layer, pooling layer, and fully connected layer.
We build our CNN model based on this type of CNN architecture. When applying CNNs in text analysis, the input has a static size and text lengths can vary greatly [18]. So, as the input, the texts from the “contact reason” field are tokenized, and all words are integer encoded. These pre-processed train data are then fed to the CNN architecture. In the embedding layer, each word is embedded into a 200-dimension vector. A one-dimensional convolution layer is then added. Filters perform convolutions on the text matrix and generate feature maps. We depict one filter region size in our experiments. The max-pooling is performed over each feature map to select the most prominent feature (the largest number). We choose max-pooling as the pooling strategy since it presents much better performance compared to average pooling in the text classification tasks [29]. Then, the max-pooling results of word embeddings are concatenated together to form a single feature vector. To prevent the built CNN model from overfitting, we add a dropout layer after the fully connected layer [30]. The final Softmax layer receives the concatenated feature vector as input and uses it to classify the texts. The main hyperparameters settings in the built CNN model are shown in the second column of Table 3.
The training and validation results are presented in Figure 3. The left figure shows the accuracy’s on the train and validation datasets. As it can be observed, the accuracy increases slowly on the training dataset after 6 epochs. The accuracy on the validation dataset tends to saturate after certain epochs, and the curve doesn’t show the trend of increasing accuracy after more epochs within the setting one. Figure 3 right plots the loss curves of the train and validation datasets over the setting epochs. The loss curve on the training dataset decreases rapidly at the beginning of 3 epochs and slowly after the 3 epochs. The loss curve on the validation dataset saturates gradually after 6 epochs, and the loss has been reduced to a minimum level at epoch 8 and then begins to increase. As observed from both figures, the curves have been basically stable after 6 epochs within the setting epochs.
Using this trained CNN model, the accuracy obtained on the test dataset is 76.56%.

4.2.2. RNN-Based Architecture

When constructing the RNN-based neural network for our task, we choose to use the pre-trained Flair model as the implementation. Flair is a natural language processing (NLP) library and builds on PyTorch which is one deep learning framework. A group of pre-trained models for the NLP tasks is provided in Flair.
We choose the pre-trained Flair model that provides text classification ability as part of the implementation of our RNN-based model [28]. The pre-trained text classification model takes word embeddings, puts them into a recurrent neural network (RNN) to obtain a text representation, and puts the text representation in the end into a linear layer to get the actual class label [28]. The list of the word embeddings from one text is passed to the RNN, and the final state of the RNN is used as the representation for the whole text. The GRU-type RNN is used in our experiments. The main hyperparameters settings used are shown in Table 3.
Figure 4 plots the accuracy and loss observed on the training and validation dataset using the RNN-based architecture. From the results, we can see that this model overfits too early after only a few epochs. The best score obtained on the testing dataset is 75.88%.

4.2.3. Transformers-Based Architecture

To construct the transformers-based neural network for our task, we chose to use the pre-trained BERT model as the implementation of transformers.
We choose a pre-trained BERT model that provides text classification ability [25,32]. With TensorFlow, we first instantiate this classification model with a pre-trained model’s configuration from a BERT Portuguese model; then, we fit the model to our dataset. We use BERTimbau Base, which is a pre-trained BERT model for Portuguese that achieves state-of-the-art performance on a number of NLP tasks. In particular, we use the base model of BERTimbau Base, which includes 12 layers [33]. The pre-processed data is then fed to the neural network. The main hyperparameter values used in the transformed-based architecture are summarized in the fourth column of Table 3.
Figure 5 depicts the accuracy and loss curves on the training and validation dataset. The loss on the validation dataset has been reduced to a minimum level at epoch 5. Using the transformers-based architecture, the accuracy score obtained on the testing dataset is 78.15%.

4.3. Fine-Tuning of the Transformers-Based Models

In our preliminary experiments, we use the default or empirical hyperparameter settings when building a prediction model. To adapt the neural network models to our task, we fine-tune the hyperparameter to our target dataset.
Since the transformers-based model presents better performance than the CNN and RNN models in the preliminary experiments, we further fine-tune the hyperparameters of the transformers-based model, aiming to achieve better results.

4.3.1. Fine-Tuning Strategy

Our fine-tuning aim is: (1) an ideal optimizer with an appropriate learning rate; (2) an ideal batch size for training on the target dataset. So, we mainly consider three hyperparameters, learning rate, optimizer, and batch size.
Since it is costly to train transformers-based models, the fine-tuning process is designed as the following:
  • Fine-tuning of learning rate (abbr. lr). During the first phase, we experiment with small epochs (e.g., 10) and range with different learning rates on the different optimizers.
  • Fine-tuning on epochs. Based on the results observed from the first phase, we analyze the accuracy and loss curves to choose the best learning rate for each optimizer. We then use that group of hyperparameters and set the epoch to a large number for further training.
  • Fine-tuning on batch size. Based on the results obtained after fine-tuning on learning rates and epochs, the final fine-tuning is done by adjusting the batch size with a set of varying values.
In such a way, we can better save the resources in training the models and shorten the time in fine-tuning. This method practically eliminates the need to repeatedly tune on large epochs every time and theoretically achieves near-optimal classification accuracy. We will describe each phase strategy in more detail in the following subsections.

4.3.2. Optimization Algorithm Choosing

The optimization algorithm (a.k.a. optimizer) is one main approach used to minimize the error when training neural network models. A number of optimizers have been researched and generally used ones include Stochastic Gradient Descent (SGD), Momentum Based Gradient Descent, Mini-Batch Gradient Descent, Nesterov Accelerated Gradient (NAG), Adaptive Gradient Algorithm (AdaGrad), Adaptive Moment Estimation (Adam), etc. [34,35,36,37]. When choosing an optimizer for fine-tuning the neural network models, the speed of convergence and the generalization performance on the new data are usually considered.
In our work, we mainly take into account two widely used optimizers in our experiments, Adaptive Moment Estimation (Adam) [34] and Stochastic Gradient Descent (SGD) [37] since these two optimizers present better performances than the others in many NLP tasks. When training neural network models, Adam is one of the most practical optimizer. Adam is an algorithm for gradient-based optimization and combines the advantages of two SGD extensions: RMSProp and Adagrad [34]. Adam can compute adaptive learning rates for different parameters individually. As a variant of Gradient Descent (GD), SGD is a computationally efficient optimization method on large-scale datasets. When doing experiments on a large-scale dataset, computations over the whole dataset are usually redundant and ineffective. SGD can do computations on a small or a randomly selected subset instead of the whole dataset [37].

4.3.3. Fine-Tuning of Learning Rate

During this phase, we explore the effect of different learning rate settings. In particular, we focus on the variants of the learning rate with each optimizer.
The setting of the learning rate plays a decisive role in the convergence of a neural network model. A too-small learning rate will make a training algorithm converge slowly, while a too-large learning rate will make the training algorithm diverge [38]. A traditional default value for the learning rate is 0.1, and we use this as a starting point on our task problem [38,39]. For each optimizer (Adam and SGD), we try learning rates within the set of {0.1, 0.01, 1 × 10 3 , 1 × 10 4 , 1 × 10 5 , 1 × 10 6 , 1 × 10 7 }. We maintain the same training procedure as the preliminary models described in Section 4.2.3 except for changing the learning rate. We use the same number of epochs (epochs = 10) as a start. The rest of the hyperparameters remain the same (see Section 4.2.3).
When varying the learning rate settings with optimizer Adam (Figure A1) and SGD (Figure A2), the training and validation curves of the train and validation dataset are depicted. The results on accuracy as well as loss values are reported.
Observing the curves learned by using Adam with different learning rates, we find that when setting the learning rate as 1 × 10 3 or bigger, the training models fails to converge (refer Appendix B Figure A1). Especially when the learning rate is set with an aggressive value of 0.1, we observe that the loss increases rather than decreases. When the value is set as 1 × 10 4 or 1 × 10 5 , the two models archive similar performance, and the models converge too early on the validation dataset. We also observe that the loss on the validation dataset increases with more epochs. When setting a rather small value of 1 × 10 7 , we observe that the loss decreases too slow compared to the other learning rates (the loss is around 2.3 after 10 epochs). When setting the learning rate as 1 × 10 6 , we observe that the results clearly show a downward trend in loss and upward in accuracy over the epochs. This sign shows the model is learning the problem and has learned the predictive skill.
So, when choosing Adam as the optimizer, the appropriate setting of learning rate is 1 × 10 6 . The loss curves of the training and validation datasets are decreasing, and they are still showing downward trends by the pre-defined epochs. This observation suggests that the prediction performance could be potentially improved with more epochs (training with more epochs than 10). The best accuracy on the test dataset achieves a score of 79.67%, obtained with the learning rate of 1 × 10 5 and Adam optimizer.
When varying the learning rate setting with optimizer SGD, we can analyze the results similarly as with the Adam optimizer (refer Appendix B Figure A2). We observe that the model fails to converge when the learning rate is set as 0.1 or 0.01; the model converges too early with a learning rate of 1 × 10 3 ; the loss decreases very slow when the learning rate is 1 × 10 5 or smaller. The best accuracy on the test dataset achieves a score of 79.72%, obtained with the learning rate of 1 × 10 3 and SGD optimizer. But the curve shows no obvious downward trend on 1 × 10 3 after 10 epochs. So, with SGD as the optimizer, the appropriate learning rate setting is 1 × 10 4 .

4.3.4. Fine-Tuning on Epochs

Based on the results observed on fine-tuning of the learning rates, we focus on adjusting the number of training epochs during this phase.
We observe that the results clearly show a downward trend in loss and upward in accuracy with the learning rate of 1 × 10 6 on Adam optimizer (see Figure A1). This trend suggests that continuing to increase the value of epochs can potentially further improve the performance of the model [40]. Although the best testing accuracy is obtained with a learning rate of 1 × 10 5 , its learning curve doesn’t show an obvious downward trend (see Figure A1). So, we choose the learning rate of 1 × 10 6 for further fine-tuning. We experiment by increasing the number of epochs from 10 to 150 and keeping the other hyperparameters the same. The training and validation curve is presented in Figure 6. The increasing trend of the loss on the validation dataset is a sign of overfitting. As presented in Figure 6, the loss curve shows the beginnings of this type of pattern after 41 epochs. This is where the model overfits the training dataset at the cost of worse performance on the validation dataset. These observations suggest that the best prediction model can be obtained after 41 epochs. The accuracy on the test dataset achieves a score of 79.96%.
Now we turn to the model trained with the SGD optimizer. As the loss decreases most quickly on SGD is when setting the learning rate to 1 × 10 4 (see Figure A2). Similar to the experiment carried out on Adam, we use SGD as the optimizer and set the learning rate to 1 × 10 4 . The results after 150 epochs are presented in Figure 7. we also observe that the runs show the beginnings of overfitting on the validation dataset after 67 epochs, and the best prediction model can be potentially obtained after 67 epochs. The accuracy on the test dataset achieves a score of 79.76%.

4.3.5. Fine-Tuning on Batch Size

As found by Smith and Le [41], small batch size can improve the performance, we further fine-tune our models by choosing different batch sizes within the set of {256, 128, 64, 32, 16}. Since the model trained with Adam optimizer and the learning rate of 1 × 10 6 is able to achieve better results than the other models after 150 epochs, the final tuning is carried out on this model.
As observed in Figure 6, the validation accuracy does not change too much and begins to drop after 41 epochs, which means the model begins to converge. So, we choose the epoch of 50 on our final tuning within the batch size set (refer Appendix B Figure A3). The model trained with a batch size of 16 is able to surpass all the other models. The accuracy on the test dataset achieves a score of 80.57%.
In summary, all the models trained during the fine-tuning process are evaluated in testing accuracy and the results are presented in Table 4.

5. Results and Discussion

We consider three deep neural network architectures in our experiments. We use these neural networks combined with different representations for words and texts. These built models and their respective testing accuracies results are presented in Table 5.
Among them, the transformers-based model using BERT-base language model achieves a result of 78.15%. The CNN-based approach achieved an accuracy of 76.56%, and the RNN-based approach achieved an accuracy of 75.88%. The transformers-based approach is shown to be better than the one using the CNN-based or RNN-based approach. We then fine-tune the transformers-based models. A better model with a higher accuracy score is obtained after we fine-tune the hyperparameters on learning rate, optimizer, number of epochs, and batch size step by step. The best fine-tuning model achieves an accuracy of 80.57% on the testing set.
We also compare the performance between the Adam and SGD optimizers using their validation accuracy distributions over the 150 epochs. In Figure 8, we present the results with a box plot using the data that both optimizers achieve with their best learning rate, where Adam has a learning rate of 1 × 10 6 and SGD has a rate of 1 × 10 4 The outliers that are smaller than the minimum for both optimizers are removed in the figure for a clear presentation. This does not affect the comparison between the distributions. We choose a learning rate of 1 × 10 6 for Adam and 1 × 10 4 for SGD because their learning curves show clear downward trends in losses and upward trends in accuracies. Also, they are able to achieve comparative performances compared to other learning rates (see Figure A1 and Figure A2).
We observe that the model with Adam optimizer achieves better top accuracy on the validation dataset than the model with SGD. Besides, when comparing Figure 6 and Figure 7, Adam has been found to attain the top accuracy performance with a smaller number of epochs necessary (epoch = 41), while SGD requires more to converge (epoch = 47).
These findings suggest that: (i) The value of the learning rate and the optimizer choice can have a strong influence on each other, in other words, the performance of the built models highly depends on the combination of the learning rate and the optimizer; (ii) With the appropriate learning rate setting, Adam outperforms SGD not only in final generalization performance but also able to converge with a smaller number of epochs.
In our previous work [17], we have explored using shallow machine learning techniques to build prediction models on the SNS24 clinical triage classification task. Four shallow machine learning algorithms were used in the previous work: linear SVM, RBF kernel SVM, Random Forest, and Multinomial Naïve Bayes [17]. For the features, we used word n-grams with word counts and TF-IDF; we also used word embeddings built with BERT [25] and Flair [27] models. The experiments carried out in this work use the same dataset and train/val/test splits. The performances between the shallow machine learning techniques and neural networks are presented in Table 6.
We can see that the transformers-based architecture achieves the highest score of all the experimented techniques, and was able to surpass the other two neural network architectures and the linear SVM algorithm, presenting the best among the four shallow machine learning techniques.

6. Analysis among Similar Clinical Symptoms

In this section, we performed an analysis of the trained neural network models when dealing with classes that have similar clinical symptoms. According to the SNS24 call center, there is some degree of overlap between clinical pathways. Although this overlap is expected, it may make cause the clinical decision making process to be more complex. Figure 9 and Figure 10 presents two example groups that includes clinical pathways with similar symptoms.
Figure 9 presents an example group that includes six clinical pathways with similar symptoms. They are Pr. da orofaringe (Oropharynx problem), Pr. deasma ou pieira (Asthma or wheezing problem), Pr. nasal (Nasal problem), Pr. respiratorio (Respiratory problem), Sindrome gripal (Flu syndrome), and Tosse (Cough).
Using the pre-trained neural network model, we built the heatmap for these clinical pathways. Taking Oropharynx problem for example, 64% of the cases are rightly classified, 4% are wrongly classified as Nasal problem, 11% as Flu syndrome, 8% as Cough, and 13% as the other clinical pathways. On the other way around, 3% Flu syndrome cases are wrongly classified as Oropharynx problem, and 12 % as Cough.
These observations suggest that Flu syndrome cases are more likely to be classified as Cough than Oropharynx problem. But on the other way, the Oropharynx problem case is more likely to be classified as Flu syndrome than Cough. Furthermore, we find that 1% of the Cough cases are wrongly classified as Oropharynx problem and 8% as Flu syndrome, which means that Cough cases are more likely to be classified as Flu syndrome.
Joining these above observations, we can deduce that it is harder to distinguish between Cough and Flu syndrome cases, and in practical application, the nurses should pay more attention to identifying these two types of clinical pathways.
Figure 10 presents an example group that includes three clinical pathways with similar symptoms. They are Diarreia (Diarrhea), Dor abdominal (Abdominal pain), and Náuseas e vômitos (Nausea and vomiting).
As can be observed, an average of 80% cases in this group can be rightly classified with the built neural network model. This means that the built model can provide useful suggestions to the SNS24 nurses when distinguish the clinical pathways that fall into this group.
We also find that for Abdominal pain, 6% are wrongly classified as Nausea and vomiting, and 3% as Diarrhea. This means that Abdominal pain is more likely to be classified as Nausea and vomiting. Similarly, when observing Nausea and vomiting, 6% are classified as Abdominal pain and 1% as Diarrhea. These findings suggest that Abdominal pain and Nausea and vomiting cases are harder to be distinguished in this group. These results provide valuable analysis for the SNS24 Health Calls center, for example, helping to do deep analysis between similar symptoms.

7. Analysis of Individual Predictions

In this section, we aim to compare the labels ascribed by the neural network model and the support vector machines predictions (model from previous work) with the original labels.
Over time, the number of different nurses working in the health line is very high, and some nurses can be biased in the selection according to their experience or other factors. For example, in the 3-month period of the collected data, 588 nurses attended calls and chose the most adequate clinical pathway according to their insight.
This analysis was done for the ten clinical pathways with most records in the dataset, selecting ten examples from each class where the “contact reason” text had no common words with the clinical pathway. This resulted in a set of 100 examples.
From the comparison between the triage given by the phone line and the two ML “experts”, we observe that both ML models agree on 78 and all three agree on 49 examples out of the 100 examples; the SVM model agrees with the original triage in 53 examples, while the NN model agrees with the original triage in 55 examples.
Table 7 resumes the agreements between experts for a sub-sample containing 20 examples (Table A2 presents the “contact reason” of each example along with the clinical pathway selected by the nurse; Table A3 presents the corresponding labels given by the ML models).
It is interesting to look at the text where both ML models agree but disagree with the triage. In example 2 both classify it as Nasal problem whereas the triage classify it as Cough. The text mentions “rhinorrhea” and no cough related symptoms. Also example 3, “Lack of appetite 6 h ago”, is classified as Nonspecific problems while triage selected Nausea and vomiting problems. In fact, there is no word in the text that gives a hint about nausea.
On the other hand, there are examples where the ML experts select pathways that have close connection to words appearing in the “contact reason”: example 5 is classified by both models as Pregnancy problem, puerperium while the Abdominal pain pathway was chosen on triage and the “contact reason” text includes the word “pregnancy”; examples 8 and 14, that include the words “vaginal” and “blood loss”, respectively, are classified by both models as Woman health problem while the pathways chosen on triage were Oropharynx problem and Urinary problem, respectively.
Nonetheless, sometimes they seem to be too biased: for example, the contact reason with text “itching starting 3 days ago, in the armpits and genital region” (example 10) is classified as Woman health problem, but there’s no body part mentioned specific to women.
Sometimes we see some ambiguity, as in example 16, which both models classify as Nausea and vomiting problems, while the triage selected Diarrhea and the text includes both the words “vomiting” and “diarrhea”.
On example 7, with text “Canker sores 1 day ago, fever 2 days ago”, the NN agrees with the triage on the pathway Oropharynx problem while NN chooses Face problem for example 12, with text “sore throat since Friday and fever 38.7”, SVM is the model that agrees with the triage on Flu syndrome while NN chooses Oropharynx problem.
Two examples where the three experts disagree have triage as Chest pain. In example 17 with text “Easy fatigue 1h ago, with irradiation”, SVM classified as Nonspecific problems while NN as Breathing problem; in example 18 with text “Flank pain since 15 days pregnant”, SVM classified as Pregnancy problem, puerperium and NN as Lower limb problems, Hip.

8. Conclusions and Future Work

SNS24 is a telephone and digital public service which provides clinical services. SNS24 plays a significant role in identifying users’ clinical situations according to their symptoms. There are a group of possible clinical algorithms defined at present, and choosing the appropriate clinical algorithm is important in each telephone triage.
We carried out experiments on a dataset containing 269,654 call records belonging to 51 classes. Each record is labeled with a clinical pathway chosen by the nurse that answered the call. Three neural network architectures were trained to produce classification models. CNN, RNN, and transformers-based approaches achieve accuracies of 76.67%, 78.03%, and 78.15%, respectively. The best performance obtained, with an accuracy of 80.57%, was reached using a fine-tuned transformers-based model. The fine-tuning strategy used showed up to be efficient to improve the model performance. Also, the experiments revealed that, when setting the appropriate learning rate, the Adam optimizer achieved better performance compared to SGD. These results suggest that using deep learning is an effective and promising approach to aid the clinical triage of phone call services.
We carried out further analysis using the fine-tuned neural network models. We used the trained model to analyze similar clinical symptoms. These results provide valuable analysis for the SNS24 Health Calls center to do deep analysis among similar symptoms. We also compared the predictions from the neural network model, the support vector machines model from our previous work, and the original labels provided by SNS24.
As future work, and since the neural network models performed well on the SNS24 clinical trial classification task, we intend to fine-tune these models with other hyperparameters besides the ones explored in this work. As a continuation of the work, we plan to build models using a much larger SNS24 dataset that includes three full years of data. To support the choices made by ML models and better understand why one clinical pathway is chosen over others, we intend to apply explainable artificial intelligence (XAI) techniques.
Pursuing the line of work from Section 7 we intend to use a new subset of examples and extend the presented comparisons to the new models being created with the three-years dataset, as well as adding the insight from a few experienced nurses from the health-line.

Author Contributions

Conceptualization and methodology, T.G. and P.Q.; software, validation and formal analysis, H.Y.; investigation, R.V. (Renata Vieira); resources, C.S.P., J.O., M.C.F., J.M., A.R.P., N.F. and C.G.; data curation, R.V. (Rute Veladas); writing, H.Y. and T.G.; visualization, H.Y.; supervision, T.G.; project administration, P.Q. and T.G.; funding acquisition, P.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by FCT (Fundação para a Ciência e a Tecnologia), I.P, within the project SNS24.Scout.IA: Aplicação de Metodologias de Inteligência Artificial e Processamento de Linguagem Natural no Serviço de Triagem, Aconselhamento e Encaminhamento do SNS24 (ref. DSAIPA/AI/0040/2019).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Authors would like to thank Javier León for helping building the experimental environments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SNS24Portugal Centro de Contacto do Serviço Nacional de Saúde
DGSDirectorate General of Health
INEMNational Medical Emergency Institute
SPMSServiços Partilhados do Ministério da Saúde
DLDeep Learning
NNNeural Network
BERTBidirectional Encoder Representations from Transformers
CNNConvolutional Neural Network
RNNRecurrent Neural Network RNN
GRUGated Recurrent Unit
NLPNatural Language Processing
SGStochastic Gradient Descent
AdamAdaptive Moment Estimation

Appendix A. Dataset: Clinical Pathways and Examples

Table A1 presents the 52 clinical pathways present in the dataset (calls from January to March 2018). For each clinical pathway, the original Portuguese name, the number, and the percentage are listed. We also translate the Portuguese name to English for the readers to have an easy understanding of the clinical pathways defined by SNS24.
Table A1. The 52 clinical pathways within the dataset collected by the SNS24 phone-line from January to March 2018 (sorted by the number of occurrences). The label is reported in the original language Portuguese and English.
Table A1. The 52 clinical pathways within the dataset collected by the SNS24 phone-line from January to March 2018 (sorted by the number of occurrences). The label is reported in the original language Portuguese and English.
Clinical Pathway (Label)Number%
Tosse (Cough)13793014.066
Síndrome Gripal (Flu Syndrome)3426612.707
Pr. por náuseas e vómitos (Nausea and vomiting problems)144535.360
Dor abdominal (Abdominal pain)143825.333
Pr. da orofaringe (Oropharynx problem)135035.007
Rash (Rash)114184.234
Pr. de alteração de temperatura corporal (Body temperature change problem)92853.443
Dor torácica (Chest pain)88053.265
Cefaleia (Migrain)83333.090
Pr. Urinário (Urinary Problem)79922.964
Diarreia (Diarrhea)79092.933
Pr. do ouvido (Ear problem)72702.696
Pr. ocular (Eye problem)69882.591
Pr. nasal (Nasal problem)54502.021
Pr. por tonturas (Problem for dizziness)53972.001
Pr. da gravidez, puerpério (Pregnancy problem, puerperium)46901.739
Pr. por lombalgia (Problem for low back pain)46721.733
Pr. na face (Face problem)45091.672
Pr. de tensão arterial (Blood pressure problem)43861.626
Pr. na cabeça e pescoço (Head and neck problem)42791.587
Pr. de saúde da mulher (Woman health problem)41601.543
Pr. de integridade cutânea (Skin integrity problem)41001.520
Pr. inespecíficos (Nonspecific problems)38301.420
Pr. de Ansiedade (Anxiety Problem)35371.312
Pr. por ingestão de substâncias tóxicas (Problem with ingestion of toxic substances)32041.188
Pr. nos membros inferiores, Tornozelo Pé (Lower limb problems, Ankle Foot)29851.107
Pr. nos membros inferiores, Anca (Lower limb problems, Hip)25540.947
Pr. respiratório (Breathing problem)23900.886
Pr. de diabetes (Diabetes problem)23840.884
Pr. de obstipação (Constipation Problem)22220.824
Pr. nos membros inferiores, Joelho (Lower limb problems, Knee)21640.802
Pr. geriátricos (Geriatric Problems)21620.802
Pr. nos membros superiores–Ombro/Clavícula/Braço (Upper limb problems–Shoulder/Collarbone/Arm)21530.798
Pr. na criança que chora (0–1 ano) (Problem in the crying child (0–1 year))19280.715
Pr. de alteração da cor das fezes (Stool color change problem)18710.694
Pr. nos membros superiores–Punho/Mão (Upper limb problem–Wrist/Hand)15210.564
Pr. de saúde do homem (Men’s health problem)14450.536
Pr. de alergias (Allergy problem)13600.504
Pr. nos dedos (Finger problems)13200.489
Pr. de desmaio ou lipotimia (Fainting problem or lipothymia)9440.350
Pr. de depressão (Depression problem)8760.325
Pr. de reação a vacinação (Vaccination reaction problems)8440.313
Emergência (Emergency)7260.269
Pr. da mama (Breast problem)7030.261
Pr. das queimaduras (Burn problems)7020.260
Pr. por corpo estranho (Inalação, Aspiração) (Foreign body problem (Inhalation, Aspiration))6570.244
Pr. na amamentação (Problem in breastfeeding)3670.136
Pr. de asma ou pieira (Asthma problem or wheezing)3330.123
Pr. no cotovelo (Elbow problems)1370.051
Pr. por sarampo (Measles problems)980.036
Pr. de adaptação em situação de crise (Crisis adaptation problems)600.022
Pr. por calor (Heat problems)40.001
1 The original texts recorded in Portuguese are translated to English for easy understanding.
Table A2 presents 20 examples of “contact reason” taken from the 10 most frequent clinical pathways. For each example, the original Portuguese text and the corresponding selected clinical pathway is presented. These were also translated to English for a better understanding.
Table A3 lists the SVM and NN predictions for each example presented in Table A2.
Table A2. Example texts of the “contact reason” field and their labels (“Clinical Pathway”).
Table A2. Example texts of the “contact reason” field and their labels (“Clinical Pathway”).
IdContact ReasonClinical Pathway
Hemoptises esta manhã
1(Hemoptysis this morning) 1Tosse  
2Rinorreia alaranjada fluoresceste ha 2 horas(Cough)
(Fluorescent orange rhinorrhea 2 h ago)
Falta de apetite há 6h
3(Lack of appetite 6 h ago)Pr. por Náuseas e Vómitos  
4Indiposição após toma de ATB há 1 dia(Pr. for Nausea and Vomiting)
(Indisposition after taking ATB for 1 day)
Dor nas costas que irradia para a face lateral esquerda há 3 dias em grávida de 5 semanas
5(Back pain radiating to the left side for 3 days in a 5-week pregnant woman)Dor Abdominal 
6Epigastralgia, ador, pirose, enfartamento desde há 4 dia(Abdominal pain)
(Epigastralgia, pain, heartburn, bloating since 4 days ago)
Aftas há 1 dia, febre há 2 dias
7(Canker sores 1 day ago, fever 2 days ago)Pr. na Orofaringe 
8Prurido na cavidade oral e a nivel vaginal sob antibioterapia.
(Itching in the oral cavity and at the vaginal level under antibiotic therapy)(Oropharynx problem)
Pediculose e tratamento para o mesmo em situação de varicela.
9(Pediculosis and treatment for the same in chickenpox situation)Rash  
10Prurido com inicio ha 3 dias, nas axilas e região genital(Rash)
(Itching starting 3 days ago, in the armpits and genital region)
Congestionamento, tosse, sub febril febre
11(Congestion, cough, sub-febrile fever)Síndrome Gripal  
12dor de garganta desde 6f e febre 38.7(Flu syndrome)
(sore throat since friday and fever 38.7)
Disuria, ardor, poliúria e desconforto e vestigios de sangue há 5horas
13(Dysuria, burning, polyuria and discomfort and traces of blood for 5 h)Pr. Urinário 
14Perdas hemáticas em moderada quantidade desde há 2 dias.(Pr. Urinary)
(Moderate blood loss for 2 days)
Dejeções de fezes moles ha 2 meses
15(Loose stools for 2 months)Diarreia 
16Episódio de vómito há 12 horas, desde então com Diarreia.(Diarrhea)
(Vomiting episode 12 h ago, since then with diarrhea)
Cansaço facil. há 1h, com irradição
17(Easy fatigue 1h ago, with irradiation)Dor Torácica  
18Dor nos flancos desde ha 15 dias em gravida(Chest pain)
(Flank pain since 15 days pregnant)
Dor peito, cabeça e ardor gastrica desde 2f sem melhoras desde segunda
19(Chest pain, head and gastric burning since monday without improvement since monday)Cefaleia  
20Cefaleias persistentes ha 2 dias, congestão nasal e tosse seca(Migrain)
(Persistent headache for 2 days, nasal congestion and dry cough)
1 The original texts recorded in Portuguese are translated to English for easy understanding.
Table A3. Examples and corresponding clinical pathways and labels predicted by SVM (Support vector machines) and NN (Neural Network).
Table A3. Examples and corresponding clinical pathways and labels predicted by SVM (Support vector machines) and NN (Neural Network).
IdClinical PathwaySVM PredictionNN Prediction
1CoughCoughCough
2CoughNasal problemNasal problem
3Nausea and vomiting problemsNonspecific problemsNonspecific problems
4Nausea and vomiting problemsAllergy problemNonspecific problems
5Abdominal painAbdominal painBody temperature change problem
6Abdominal painAbdominal painAbdominal pain
7Oropharynx problemOropharynx problemOropharynx problem
8Oropharynx problemFace problemOropharynx problem
9RashRashRash
10RashWoman health problemWoman health problem
11Flu syndromeFlu syndromeFlu syndrome
12Flu syndromeFlu syndromeOropharynx problem
13Urinary ProblemUrinary ProblemUrinary Problem
14Urinary ProblemWoman health problemWoman health problem
15DiarrheaDiarrheaDiarrhea
16DiarrheaNausea and vomiting problemsNausea and vomiting problems
17Chest painNonspecific problemsBreathing problem
18Chest painPregnancy problem, puerperiumLower limb problems, Hip
19MigrainChest painChest pain
20MigrainCoughFlu syndrome

Appendix B. Neural Networks: Training and Validations Curves

Figure A1 and Figure A2 present the training and validation curves when varying the learning rate settings with optimizer Adam and SGD, respectively. Learning rates are varied within the set of {0.1, 0.01, 1 × 10 3 , 1 × 10 4 , 1 × 10 5 , 1 × 10 6 , 1 × 10 7 } for each optimizer. We present the results on accuracy and loss values. The training procedure and the hyperparameters settings remain the same as the preliminary models (see Section 4.2.3). The learning rates are experimented within the defined set.
Figure A3 plots the fine-tuning of the models by choosing different batch sizes. The final fine-tuned models are trained by using Adam optimizer, the learning rate of 1 × 10 6 , the epoch of 50, and varying the batch sizes within the set of 256, 128, 64, 32, 16.
Figure A1. Training and validation performance with different learning rates on Adam optimizer over 10 epochs using the transformers-based model.
Figure A1. Training and validation performance with different learning rates on Adam optimizer over 10 epochs using the transformers-based model.
Futureinternet 14 00130 g0a1
Figure A2. Training and validation performance with different learning rates on SGD optimizer over 10 epochs using the transformers-based model.
Figure A2. Training and validation performance with different learning rates on SGD optimizer over 10 epochs using the transformers-based model.
Futureinternet 14 00130 g0a2
Figure A3. Performance of the improved transformers-based models. The models use the Adam optimizer and a learning rate of 1 × 10 6 over 50 epochs, varying different batch sizes. The other hyperparameters are set the same as the model built in the preliminary experiments.
Figure A3. Performance of the improved transformers-based models. The models use the Adam optimizer and a learning rate of 1 × 10 6 over 50 epochs, varying different batch sizes. The other hyperparameters are set the same as the model built in the preliminary experiments.
Futureinternet 14 00130 g0a3

References

  1. Mackway-Jones, K.; Marsden, J.; Windle, J. Emergency Triage: Manchester Triage Group; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  2. Rajkomar, A.; Dean, J.; Kohane, I. Machine learning in medicine. N. Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef] [PubMed]
  3. Spiga, O.; Cicaloni, V.; Dimitri, G.M.; Pettini, F.; Braconi, D.; Bernini, A.; Santucci, A. Machine learning application for patient stratification and phenotype/genotype investigation in a rare disease. Briefings Bioinform. 2021, 22, bbaa434. [Google Scholar] [CrossRef]
  4. Sidey-Gibbons, J.A.; Sidey-Gibbons, C.J. Machine learning in medicine: A practical introduction. BMC Med. Res. Methodol. 2019, 19, 64. [Google Scholar] [CrossRef] [Green Version]
  5. Shin, B.; Chokshi, F.H.; Lee, T.; Choi, J.D. Classification of radiology reports using neural attention models. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 4363–4370. [Google Scholar]
  6. Wu, H.; Wang, M.D. Infer cause of death for population health using convolutional neural network. In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Boston, MA, USA, 20–23 August 2017; pp. 526–535. [Google Scholar]
  7. Mullenbach, J.; Wiegreffe, S.; Duke, J.; Sun, J.; Eisenstein, J. Explainable prediction of medical codes from clinical text. arXiv 2018, arXiv:1802.05695. [Google Scholar]
  8. Hughes, M.; Li, I.; Kotoulas, S.; Suzumura, T. Medical text classification using convolutional neural networks. Stud. Health Technol. Inform. 2017, 235, 246–250. [Google Scholar]
  9. Baker, S.; Korhonen, A.L.; Pyysalo, S. Cancer Hallmark Text Classification Using Convolutional Neural Networks; The COLING 2016 Organizing Committee: Osaka, Japan, 2016. [Google Scholar]
  10. Zhou, X.; Li, Y.; Liang, W. CNN-RNN based intelligent recommendation for online medical pre-diagnosis support. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 912–921. [Google Scholar] [CrossRef] [PubMed]
  11. Gao, X.; Xu, X.; Li, D. Accuracy Analysis of Triage Recommendation Based on CNN, RNN and RCNN Models. In Proceedings of the 2021 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2021; pp. 1323–1327. [Google Scholar] [CrossRef]
  12. Gao, S.; Alawad, M.; Young, M.T.; Gounley, J.; Schaefferkoetter, N.; Yoon, H.J.; Wu, X.C.; Durbin, E.B.; Doherty, J.; Stroup, A.; et al. Limitations of Transformers on Clinical Text Classification. IEEE J. Biomed. Health Inform. 2021, 25, 3596–3607. [Google Scholar] [CrossRef]
  13. Behera, B.; Kumaravelan, G.; Kumar, P. Performance evaluation of deep learning algorithms in biomedical document classification. In Proceedings of the 2019 11th International Conference on Advanced Computing (ICoAC), Chennai, India, 18–20 December 2019; pp. 220–224. [Google Scholar]
  14. Al-Garadi, M.A.; Yang, Y.C.; Cai, H.; Ruan, Y.; O’Connor, K.; Graciela, G.H.; Perrone, J.; Sarker, A. Text classification models for the automatic detection of nonmedical prescription medication use from social media. BMC Med. Inform. Decis. Mak. 2021, 21, 27. [Google Scholar] [CrossRef] [PubMed]
  15. Mascio, A.; Kraljevic, Z.; Bean, D.; Dobson, R.; Stewart, R.; Bendayan, R.; Roberts, A. Comparative analysis of text classification approaches in electronic health records. arXiv 2020, arXiv:2005.06624. [Google Scholar]
  16. Flores, C.A.; Figueroa, R.L.; Pezoa, J.E. Active Learning for Biomedical Text Classification Based on Automatically Generated Regular Expressions. IEEE Access 2021, 9, 38767–38777. [Google Scholar] [CrossRef]
  17. Veladas, R.; Yang, H.; Quaresma, P.; Gonçalves, T.; Vieira, R.; Sousa Pinto, C.; Martins, J.P.; Oliveira, J.; Cortes Ferreira, M. Aiding Clinical Triage with Text Classification. In EPIA Conference on Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2021; pp. 83–96. [Google Scholar]
  18. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar] [CrossRef] [Green Version]
  19. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  20. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  21. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  23. Li, Y.; Yang, T. Word Embedding for Understanding Natural Language: A Survey. In Studies in Big Data; Springer: Cham, Switzerland, 2017; Volume 26. [Google Scholar] [CrossRef]
  24. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; Volume 1, pp. 2227–2237. [Google Scholar] [CrossRef] [Green Version]
  25. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  26. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. Xlnet: Generalized autoregressive pretraining for language understanding. Adv. Neural Inf. Process. Syst. 2019, 32, 5753–5763. [Google Scholar]
  27. Akbik, A.; Blythe, D.; Vollgraf, R. Contextual String Embeddings for Sequence Labeling. In Proceedings of the COLING 2018, 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, 20–28 August 2018; pp. 1638–1649. [Google Scholar]
  28. Akbik, A.; Bergmann, T.; Blythe, D.; Rasul, K.; Schweter, S.; Vollgraf, R. FLAIR: An easy-to-use framework for state-of-the-art NLP. In Proceedings of the NAACL 2019, 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations), Minneapolis, MN, USA, 2–7 June 2019; pp. 54–59. [Google Scholar]
  29. Zhang, Y.; Wallace, B. A sensitivity analysis of (and practitioners’ guide to) convolutional neural networks for sentence classification. arXiv 2015, arXiv:1510.03820. [Google Scholar]
  30. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  31. Hendrycks, D.; Gimpel, K. Bridging Nonlinearities and Stochastic Regularizers with Gaussian Error Linear Units. 2016. Available online: https://www.researchgate.net/publication/304506026_Bridging_Nonlinearities_and_Stochastic_Regularizers_with_Gaussian_Error_Linear_Units (accessed on 2 April 2022).
  32. Alatawi, H.S.; Alhothali, A.M.; Moria, K.M. Detecting White Supremacist Hate Speech using Domain Specific Word Embedding with Deep Learning and BERT. arXiv 2020, arXiv:2010.00357. [Google Scholar] [CrossRef]
  33. Souza, F.; Nogueira, R.; Lotufo, R. BERTimbau: Pretrained BERT models for Brazilian Portuguese. In Proceedings of the 9th Brazilian Conference on Intelligent Systems, BRACIS, Rio Grande do Sul, Brazil, 20–23 October 2020. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Choi, D.; Shallue, C.J.; Nado, Z.; Lee, J.; Maddison, C.J.; Dahl, G.E. On empirical comparisons of optimizers for deep learning. arXiv 2019, arXiv:1910.05446. [Google Scholar]
  36. Vani, S.; Rao, T.M. An experimental approach towards the performance assessment of various optimizers on convolutional neural network. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; pp. 331–336. [Google Scholar]
  37. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the COMPSTAT’2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 177–186. [Google Scholar]
  38. Bengio, Y. Practical recommendations for gradient-based training of deep architectures. In Neural Networks: Tricks Trade; Springer: Berlin/Heidelberg, Germany, 2012; pp. 437–478. [Google Scholar]
  39. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  40. Smith, L.N. Cyclical learning rates for training neural networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 464–472. [Google Scholar]
  41. Smith, S.L.; Le, Q.V. A bayesian perspective on generalization and stochastic gradient descent. arXiv 2017, arXiv:1710.06451. [Google Scholar]
Figure 1. Applying deep learning approaches on SNS24 clinical trial classification. The clinical texts which are recorded by SNS24 nurses are pre-processed and as the input of the neural network models. Using the designed deep learning architectures, neural network models are trained to predict clinical pathways and provide suggestions to the nurses.
Figure 1. Applying deep learning approaches on SNS24 clinical trial classification. The clinical texts which are recorded by SNS24 nurses are pre-processed and as the input of the neural network models. Using the designed deep learning architectures, neural network models are trained to predict clinical pathways and provide suggestions to the nurses.
Futureinternet 14 00130 g001
Figure 2. Using neural network architectures to build prediction models for the SNS24 clinical trial classification task. The input layer, hidden layers, and output layer are designed for each of the three neural network architectures accordingly. In the input layer, the “contact reason” and “clinical pathway” are fed to the neural network. In the hidden layers, the pre-trained word embeddings or language models are selected for each architecture accordingly. The output layer is designed to do the predictions.
Figure 2. Using neural network architectures to build prediction models for the SNS24 clinical trial classification task. The input layer, hidden layers, and output layer are designed for each of the three neural network architectures accordingly. In the input layer, the “contact reason” and “clinical pathway” are fed to the neural network. In the hidden layers, the pre-trained word embeddings or language models are selected for each architecture accordingly. The output layer is designed to do the predictions.
Futureinternet 14 00130 g002
Figure 3. Training and validation results using the CNN-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Figure 3. Training and validation results using the CNN-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Futureinternet 14 00130 g003
Figure 4. Training and validation results using the RNN-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Figure 4. Training and validation results using the RNN-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Futureinternet 14 00130 g004
Figure 5. Training and validation results using the transformers-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Figure 5. Training and validation results using the transformers-based architecture. (Left): Accuracy vs. epochs; (Right): Loss vs. epochs.
Futureinternet 14 00130 g005
Figure 6. Performance of the improved transformers-base model. This model uses the Adam optimizer and a learning rate of 1 × 10 6 over 150 epochs. The other hyperparameters are set the same as the model built in the preliminary experiments.
Figure 6. Performance of the improved transformers-base model. This model uses the Adam optimizer and a learning rate of 1 × 10 6 over 150 epochs. The other hyperparameters are set the same as the model built in the preliminary experiments.
Futureinternet 14 00130 g006
Figure 7. Performance of the improved transformers-base model. This model uses an SGD optimizer and a learning rate of 1 × 10 4 over 150 epochs. The other hyperparameters are set the same as the model built in the preliminary experiments.
Figure 7. Performance of the improved transformers-base model. This model uses an SGD optimizer and a learning rate of 1 × 10 4 over 150 epochs. The other hyperparameters are set the same as the model built in the preliminary experiments.
Futureinternet 14 00130 g007
Figure 8. The validation accuracy comparison between SGD and Adam over 150 epochs using transformers-based architecture.
Figure 8. The validation accuracy comparison between SGD and Adam over 150 epochs using transformers-based architecture.
Futureinternet 14 00130 g008
Figure 9. Analysis among clinical pathways with similar clinical symptoms using the trained neural network model. This group includes six clinical pathways: Pr. da orofaringe (Oropharynx problem), Pr. deasma ou pieira (Asthma or wheezing problem), Pr. nasal (Nasal problem), Pr. respiratorio (Respiratory problem), Sindrome gripal (Flu syndrome), Tosse (Cough).
Figure 9. Analysis among clinical pathways with similar clinical symptoms using the trained neural network model. This group includes six clinical pathways: Pr. da orofaringe (Oropharynx problem), Pr. deasma ou pieira (Asthma or wheezing problem), Pr. nasal (Nasal problem), Pr. respiratorio (Respiratory problem), Sindrome gripal (Flu syndrome), Tosse (Cough).
Futureinternet 14 00130 g009
Figure 10. Analysis among clinical pathways with similar clinical symptoms using the trained neural network model. This group includes three clinical pathways: Diarreia (Diarrhea), Dor abdominal (Abdominal pain), and Náuseas e vômitos (Nausea and vomiting).
Figure 10. Analysis among clinical pathways with similar clinical symptoms using the trained neural network model. This group includes three clinical pathways: Diarreia (Diarrhea), Dor abdominal (Abdominal pain), and Náuseas e vômitos (Nausea and vomiting).
Futureinternet 14 00130 g010
Table 1. The 18 fields recorded in the original dataset. The fields are listed in the original language Portuguese and English.
Table 1. The 18 fields recorded in the original dataset. The fields are listed in the original language Portuguese and English.
IdField (Portuguese)Field (English)
1ID encontromeeting ID
2numero snssns number
3centro saude descricaohealth center description
4data iniciostart date
5data disposicao finalfinal disposition date
6data nascimentobirth date
7idadeage
8generogender
9relacao com utenterelationship with user
10intencao inicialinitial intention
11motivo contactocontact reason
12ultimo algoritmolast algorithm
13disposicao finalfinal disposition
14comentarioscomments
15unidade saude encaminhamentoreferral health unit
16seguimentofollow-up
17tipologia interacaointeraction typology
18canalchannel
Table 2. The statistics of the original and experimental dataset.
Table 2. The statistics of the original and experimental dataset.
DatasetCollecting PeriodRecordsFieldsClasses
Original1 January 2018–31 March 2018269,65818 fields52
Experimental1 January 2018–31 March 2018269,654“contact reason” field51
Table 3. Hyperparameter settings in the neural network models for different architectures. The first column lists the main hyperparameters used in all three neural network architectures, and the hyperparameters that are not used or not manually set in our experiments in one architecture are marked with “-” in the table. CCEE stands for categorical cross entropy error.
Table 3. Hyperparameter settings in the neural network models for different architectures. The first column lists the main hyperparameters used in all three neural network architectures, and the hyperparameters that are not used or not manually set in our experiments in one architecture are marked with “-” in the table. CCEE stands for categorical cross entropy error.
HyperparametersCNN-BasedRNN-BasedTransformers-Based
Hidden layers4312
Hidden size-512768
Filter size3--
Feature map128--
Pooling strategy1-max pooling--
Activation functionReLU, Softmax-Gelu [31]
Regularization strategyDropoutDropoutDropout
Dropout rate0.10.10.1
Batch size128128128
Patience-5-
Anneal factor-0.5-
OptimizerAdamAdamAdam
Learning rate 1 × 10 4 1 × 10 4 1 × 10 4
Loss functionCCEECCEECCEE
Epoch101010
Table 4. Model results obtained during the fine-tuning process. All the models use the transformers-based architecture. The hyperparameters that are not listed in this table use the same settings presented in Table 3.
Table 4. Model results obtained during the fine-tuning process. All the models use the transformers-based architecture. The hyperparameters that are not listed in this table use the same settings presented in Table 3.
OptimizerLearning RateEpochsBatch SizeTesting Accuracy
Adam0.1101286.57%
Adam0.01101289.05%
Adam 1 × 10 3 1012814.02%
Adam 1 × 10 4 1012878.01%
Adam 1 × 10 5 1012879.67%
Adam 1 × 10 6 1012876.57%
Adam 1 × 10 7 1012852.32%
SGD0.11012812.12%
SGD0.011012873.76%
SGD 1 × 10 3 1012879.72%
SGD 1 × 10 4 1012875.01%
SGD 1 × 10 5 1012838.89%
SGD 1 × 10 6 1012815.01%
SGD 1 × 10 7 101286.26%
Adam 1 × 10 6 15012879.96%
SGD 1 × 10 4 15012879.76%
Adam 1 × 10 6 501680.57%
Adam 1 × 10 6 503280.19%
Adam 1 × 10 6 506480.20%
Adam 1 × 10 6 5012880.25%
Adam 1 × 10 6 5025680.08%
Table 5. Results comparison among the three deep learning architectures.
Table 5. Results comparison among the three deep learning architectures.
Deep Learning ArchitectureFine-Tuned HyperparametersTesting Accuracy
CNN-based architecture-76.56%
RNN-based architecture-75.88%
Transformers-based architecture-78.15%
learning rate79.67% (Adam, lr =  1 × 10 5 )
learning rate79.72% (SGD, lr =  1 × 10 3 )
epochs79.96% (Adam, epochs = 150)
epochs79.76% (SGD, epochs = 150)
batch size80.57%
Table 6. Comparison between the neural network and shallow machine learning techniques on the SNS24 clinical triage classification task.
Table 6. Comparison between the neural network and shallow machine learning techniques on the SNS24 clinical triage classification task.
Machine Learning TechniquesArchitecture/AlgorithmTesting Accuracy
Neural networkCNN-based architecture76.56%
RNN-based architecture75.88%
Transformers-based architecture80.57%
Shallow machine learningLinear SVM77.96%
RBF SVM76.97%
Random Forest74.93%
Multinomial NB66.26%
Table 7. Agreements between SNS24 triage, neural network, and SVM experts for 20 examples. The Id column indicates the texts that the experts agreed with, and the referring text can be found with the Id in Table A2. The Count column calculates the total number of agreements among the experts.
Table 7. Agreements between SNS24 triage, neural network, and SVM experts for 20 examples. The Id column indicates the texts that the experts agreed with, and the referring text can be found with the Id in Table A2. The Count column calculates the total number of agreements among the experts.
AgreementIdCount
all experts1, 6, 9, 11, 13, 156
triage, NN71
triage, SVM121
SVM, NN2, 3, 5, 8, 10, 14, 16, 198
none4, 17, 18, 204
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, H.; Gonçalves, T.; Quaresma, P.; Vieira, R.; Veladas, R.; Pinto, C.S.; Oliveira, J.; Ferreira, M.C.; Morais, J.; Pereira, A.R.; et al. Clinical Trial Classification of SNS24 Calls with Neural Networks. Future Internet 2022, 14, 130. https://doi.org/10.3390/fi14050130

AMA Style

Yang H, Gonçalves T, Quaresma P, Vieira R, Veladas R, Pinto CS, Oliveira J, Ferreira MC, Morais J, Pereira AR, et al. Clinical Trial Classification of SNS24 Calls with Neural Networks. Future Internet. 2022; 14(5):130. https://doi.org/10.3390/fi14050130

Chicago/Turabian Style

Yang, Hua, Teresa Gonçalves, Paulo Quaresma, Renata Vieira, Rute Veladas, Cátia Sousa Pinto, João Oliveira, Maria Cortes Ferreira, Jéssica Morais, Ana Raquel Pereira, and et al. 2022. "Clinical Trial Classification of SNS24 Calls with Neural Networks" Future Internet 14, no. 5: 130. https://doi.org/10.3390/fi14050130

APA Style

Yang, H., Gonçalves, T., Quaresma, P., Vieira, R., Veladas, R., Pinto, C. S., Oliveira, J., Ferreira, M. C., Morais, J., Pereira, A. R., Fernandes, N., & Gonçalves, C. (2022). Clinical Trial Classification of SNS24 Calls with Neural Networks. Future Internet, 14(5), 130. https://doi.org/10.3390/fi14050130

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop