Next Article in Journal
A Consortium Blockchain-Based Secure and Trusted Electronic Portfolio Management Scheme
Next Article in Special Issue
Analysis of Spatially Distributed Data in Internet of Things in the Environmental Context
Previous Article in Journal
Concentration of Microparticles Using Flexural Acoustic Wave in Sessile Droplets
Previous Article in Special Issue
Hybrid Selection Method of Feature Variables and Prediction Modeling for Municipal Solid Waste Incinerator Temperature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition

School of Information & Communication Engineering, Beijing Information Science and Technology University, Beijing 100025, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1270; https://doi.org/10.3390/s22031270
Submission received: 11 November 2021 / Revised: 17 January 2022 / Accepted: 20 January 2022 / Published: 8 February 2022

Abstract

:
The ‘intention’ classification of a user question is an important element of a task-engine driven chatbot. The essence of a user question’s intention understanding is the text classification. The transfer learning, such as BERT (Bidirectional Encoder Representations from Transformers) and ERNIE (Enhanced Representation through Knowledge Integration), has put the text classification task into a new level, but the BERT and ERNIE model are difficult to support high QPS (queries per second) intelligent dialogue systems due to computational performance issues. In reality, the simple classification model usually shows a high computational performance, but they are limited by low accuracy. In this paper, we use knowledge of the ERNIE model to distill the FastText model; the ERNIE model works as a teacher model to predict the massive online unlabeled data for data enhancement, and then guides the training of the student model of FastText with better computational efficiency. The FastText model is distilled by the ERNIE model in chatbot intention classification. This not only guarantees the superiority of its original computational performance, but also the intention classification accuracy has been significantly improved.

1. Introduction

Chatbots have become more widely used in industries and smart cities with the rapid development of Artificial Intelligence. Chatbots are generally divided into either open-domain and closed-domain. In an open context, users can chat with chatbots on any topic such as entertainment, emotional companionship, and socialization, which is known as Chitchat bots. In a closed context, you can only talk about the topic set by the robot, which belongs to the closed-domain type of chatbots. In addition, closed-domain chatbots are designed to accomplish a certain task, such as customer support and shopping guidance, which is known as a task-oriented chatbot. For example, Ali Xiaomi [1] is an Intelligent Personal Assistant in the field of e-commerce; based on Ali’s massive consumption and business data, combined with the needs of online and offline life scenarios, it provides intelligent shopping guides, services, and assistants with an anthropomorphic interactive business experience in a smart manual mode. An open-domain chatbot, such as Siri from Apple or Xiaobing [2] from Microsoft, can provide a 24 h personal assistant for users.
Intention classification is essential to the performance of a task-oriented chatbot. The user’s intention determines the success of the subsequent chatbot process, as shown in Figure 1. The core of intention recognition is text classification, which can be regarded as a traditional problem in the field of natural language processing. Traditional machine learning methods for text classification include the support vector machine (SVM) [3]; however, it was limited by low accuracy for the text classification task. In recent years, with the rapid development of deep learning technology, neural networks have been applied to text classification technology. Convolution neural network (CNN) has achieved great success in the field of images [4]. ColloBERT applied CNN to natural language tasks. The TextCNN [5] applied CNN to text classification with effective results. Since CNN cannot capture long-distance dependence, and the Recurrent Neural Network (RNN) [6] can handle time series features, RNN has shown advantages in natural language processing and feature extraction. To solve the gradient disappearance problem of RNN in the training process, an RNN-based variant of the Long-Short Term Memory (LSTM) network and GRU (Gate Recurrent Unit) network structure [7] have been proposed. Additionally, there is an advance for multivariate time series features extraction [8].
RNN-based models are limited by performance issues, and it is difficult to have beneficial real-time applications, such as high-concurrency scenarios in the field of chatbots. Yang et al. [9] discussed the application of deep learning in the intention classification of chatbots. Bahdanau et al. in 2015 [10] developed an attention mechanism, while Vaswani et al. in 2017 [11] proposed a transformer as seq2seq model with parallelized attentions. In 2017, Mikolov et al. [12] proposed the FastText model; the network architecture of FastText is very similar to CBOW (Continuous Bag-of-Words) [13], and FastText is mainly used for text classification and displays positive computational efficiency. Peters et al. [14] generalizes traditional word embedding research ELMo (Embedding from Language Model) along a different dimension in 2018, and they suggested extracting context sensitive features and applying pre-training tasks to text classification tasks.
Fine-tuning based on pre-training for a language model has been developed for text generation tasks (e.g., GPT (Radford et al., 2018) [15]; Google proposed the BERT (Bidirectional Encoder Representations from Transformers) model in 2018 [16]. The BERT model is based on a self-attention model architecture. The successful application of BERT has put the natural language text classification task into a new level.
The BERT model mainly focuses on words or more fine-grained cloze learning and does not make full use of the lexical structure in the training data. Li [17] used the application of the BERT model in text classification at the sub-word level. Baidu proposed the ERNIE (Enhanced Representation through Knowledge Integration) model in 2018 [18]. Compared with BERT, ERNIE has improved two masking strategies; the first is the phrase-based masking strategy, and the second is the entity (name, location, organization, product, etc.) based masking strategy. Although the BERT and ERNIE models can achieve positive results in various tasks in the field of natural language processing, it is difficult for scientific research teams in universities as well as in small- and medium-sized enterprises to pre-train BERT from the base model due to the large size of the model, the large amount of data, and the large number of resources required for training. In the field of computer vision, Hinton et al. [19] proposed a training method based on the “teacher-student network idea” in 2015. The basic principle was to transfer the features learned from the “teacher” network with a complex network structure and strong learning ability to the “student” network with a small number of parameters and weak learning ability. Sau et al. [20] used a noise-based regularizer to train the student from the teacher for a higher training efficiency in 2016. Liao et al. [21] tried to use the BERT model for distillation learning in LSTM and CNN. Sun et al. [22] proposed Patient Knowledge Distillation for the BERT Model.
In e-commerce, it is usually computationally expensive when the chatbot handles a large number of real customers online; the simple intention classification model can show a beneficial computational performance but is limited to the low classification accuracy. To improve the performance of the small-scale model, the knowledge obtained by the “teacher” model can guide the training of the original small-scale model to achieve better results. In this study, we develop a chatbot for after-sale service in e-commerce to help manage the high computational price issue. The pre-trained ERNIE model will distill the FastText model to increase the efficiency and maintain the classification accuracy. Through the knowledge distillation method, the teacher model ERNIE is used to predict intention labels on a large amount of online data, and the student model FastText can be distilled through the enhanced data of large-scale corpus from the teacher model. With these large amounts of unlabeled text, the knowledge in the teacher model ERNIE can be better transferred to the student model FastText. In addition, several different text classification models, such as LSTM and BERT, are performed to the same task for comparison.

2. Pre-Trained Model

Pre-training models have long been used in the field of computer vision. The transfer learning is pre-training a neural network model on a known task, and then performing fine-tuning using the trained neural network as the basis to a new purpose-specific model. Recently, in the field of natural language processing (NLP), the use of language model pre-training methods has achieved beneficial improvements on multiple NLP tasks. These tasks include natural text classification, sentence comprehension, and other sentence-level tasks.
There are two strategies for applying pre-trained language features to tasks: feature-based and fine-tuning. For example, ELMo uses pre-trained features as additional features; a pre-trained neural network produces word embeddings, which are subsequently used as a feature on NLP models. Fine-tuning is to apply the learning ideas of transfer learning to the deep learning language model. The network model and parameter weights are determined mainly through pre-training. In practical applications, the pre-trained model is selectively loaded and then a new data set is added to retrain the model; this is called the basic step of fine-tuning, such as BERT.
BERT makes use of Transformer, an attention mechanism that learns contextual relations between words (or sub-words) in a text. The BERT model proposed a new pre-training target: the masked language model (MLM). The models structure is shown in Figure 2. MLM randomly selects various words to mask, with the purpose to predict the masked portion based on the context non-masked words in the sequence, the MLM can access future context information due to the bidirectional feature provided by the Transformer architecture. In addition, BERT performs the next sentence prediction task, and the model receives pairs of sentences as input and learns to predict if the second sentence in the pair is the subsequent sentence in the original corpus.

3. Knowledge Distillation

The knowledge distillation is teaching a smaller model that is more suitable for reasoning by using an already trained network; the process by which is also known as model compression. Knowledge distillation introduces the soft target of the “teacher” model with more complex reasoning ability to teach the “student” network with small parameters and weak learning ability, as shown in Figure 3. The soft label is the probability vector output for large-scale samples, and the probability vector has more information compared with the one-hot label for text classification task. By passing the soft label information of the large model to the student model, the learning ability of the student model can be improved.
Due to the large-scale softmax output, the value distributed in the correct position will be large, while the value of other positions will be small. It is often more effective to directly use the probability distribution of the teacher network to learn the student network. It is necessary to improve the softmax function, which is called a heating technique.
Original softmax function:
y i = exp ( x i ) j = 1 N exp ( x j )
where, yi is the predicted probability that the document belongs to in the j-th class, xi denotes the elements of the input vector to the softmax function, and N represents the number of classes in the multi-class classifier. The term on the bottom of the formula is the normalization term.
Improved softmax function:
y i = exp ( x i T ) j = 1 N exp ( x j T )
where T is the temperature with a fixed value.

4. Teacher Model (ERNIE)

Compared with BERT, ERNIE 1.0 has improved two masking strategies, the first based on the phrase, and the second based on the entity (name, location, organization, product, etc.), such as Beijing. In ERNIE, the phrase or entity composed of multiple words is regarded as a unified unit. Compared with the word-based mask of BERT, ERNIE is trained through the mask of the entity unit. In contrast to directly mapping knowledge queries into vectors and then adding them up directly, ERNIE can potentially learn knowledge dependence and longer semantic dependence through a unified mask to make the model more generalized, as shown in Figure 4.
In addition, ERNIE 2.0 builds lexical level, grammatical level, and semantic level pre-training tasks based on ERNIE1.0 and accumulates new knowledge through multi-tasking and continuous learning. In this task, ERNIE1.0 base is used as the teacher model.

5. Student Model (FastText)

FastText’s network architecture is very similar to CBOW; it is mainly used for text classification compared to CBOW. For text classification of short text, short text sentences are represented by a “bag of words” model, and then linear classifiers are used for text classification. FastText shows a positive computational performance in text classification tasks based on its network structure. The input layer of the FastText network model (Figure 5) is the word embedding vector of the word sequence of a sentence or a paragraph of text.
These feature vectors are averaged to generate text vectors. The output layer uses hierarchical softmax, which uses the frequency of the classification label to make the leaf nodes of HuffmanTree. Each node is associated with a probability, and the probability value is equal to the product of the probability from the root node to the leaf nodes, which belong to this class. If the depth of the leaf nodes is l + 1 , their parent node distribution is n 1 , n 2 , , n l .
Its probability can be expressed as:
P ( n l + 1 ) = i = 1 l P ( n i )
where, P(ni) denotes the conditional probability of the next word in a sequence.
To optimize the calculation, the hierarchical softmax model is adopted. The Huffman Tree is constructed by using the occurrence frequency of the term or classification label, which is used as the leaf node of the Huffman Tree. The input of CBOW is the term in the context window, and the output layer corresponds to the term with the highest probability. The input of FastText is the content of the entire sentence, including term and n = gram. The output layer corresponds to the classified label. The purpose of CBOW is to obtain the required word vector through repeated iterations. FastText traverses all nodes of the tree and finds the classification label with the highest probability.

6. Experiment and Result Analysis

It mainly used the after-sales intelligent customer service Q&A of real e-commerce as the corpus, focusing on 16 task-oriented intention recognitions. The training corpus was manually labeled, with 120,000 corpora, as shown in Table 1. In the 120,000 corpora, there were 15 customer business intentions, which mainly included the return of goods, urging delivery, request of a replacement, application for an exchange, etc. The 16th intention was other categories (non-business intentions), which mainly included chitchat-related corpus. This corpus had an average of 7000 samples per business category. In addition, there were about 10,000 test samples as the test set.
In the experiment, the different models were used to learn the training corpus. The specific training parameters of BERT and ERNIE are shown in Table 2, with a total of 12 layers and a hidden layer of 768 dimensions. The 12-head model was used (Table 2), with a total of 110 million parameters. Next, the optimized different models were used to evaluate the test data separately, as shown in Figure 6. The ERNIE model archived best the performance of accuracy with increments of the epoch.
Next, we used the ERNIE base model as the teacher model, while the FastText model was used as the student model for training with different dimensions. The size of ngram was set to 2 (Table 3).
One million online real customer service chat corpora were used for distillation. The teacher model was used to obtain the tags, and next the student model FastText was distilled. We conducted several FastText experiments with different dimensions, as shown in Figure 7, where the accuracy increased with larger dimension numbers. The best performance of accuracy reached 0.95, which did not surpass the teacher model ERNIE in its accuracy of 0.957. In addition, the experiments of different word-ngrams was also conducted, as shown in Figure 8. This experiment showed which word-ngrams reached the best accuracy performance of 0.951, of which there were two.
In addition, the LSTM and BERT algorithm models were used for intention classification on the same corpus for comparison with the same training and test corpus. The results of Bi-LSTM, BERT base, ERNIE base, FastText, and FastText (ERNIE distillation) are shown in Table 4.
It can be seen from the experimental results that the FastText model has the lowest accuracy before distillation, but it shows the best computational performance. For the Chinese intention classification problem of e-commerce customer service, the ERNIE (base) model has a significant improvement in accuracy when compared with the BERT (base) model. There is not much difference in the average time consumption of online services between these two models. FastText distilled by ERNIE (base) has a significant improvement in accuracy, 95%, with a beneficial computational performance.
The model complexity and the number of parameters of FastText have no advantages over the ERNIE model, but it has obvious advantages in computational performance and has a beneficial performance of accuracy after distillation.

7. Conclusions

This paper applied the teacher-student distillation model, as well as proposed the teacher model based on ERNIE. Further, it applied it to the student model based on FastText in its intention recognition task of a chatbot for after-sale services in e-commerce. Compared with traditional models, the ERNIE model based on transfer learning had a beneficial advantage in the task of intelligent dialogue intention classifications. However, the network structure of the ERNIE model was complex, and it was difficult to meet the online performance requirements under the condition of high concurrent qps in large-scale e-commerce scenarios. In addition, GPU was expensive and had no cost advantage. Therefore, the labeling inference of the ERNIE model was performed on one million online corpora, and then the distillation was carried out on the student model FastText by using the one million online labeled corpora labeled by teach model of ERNIE.
Experiments on real chatbot corpus demonstrated that the ERNIE model outperformed the other models, such as LSTM, FastText, and BERT. The intention recognition accuracy of the ERNIE model could reach 95.7% in 16 tasks of 100,000 intelligent dialogue tasks, which was higher than LSTM and BERT in the same task. By using ERNIE as the teacher model and distilling the student model based on FastText, the intention recognition accuracy was improved by about 0.94%, compared with the native FastText model. The performance of FastText could reach 0.1 ms, which could meet the online requirements of high concurrency.

Author Contributions

Conceptualization, S.G.; methodology, S.G.; software, S.G.; validation, S.G., Q.W.; formal analysis, S.G.; investigation, S.G.; resources, S.G.; data curation, S.G.; writing—original draft preparation, S.G.; writing—review and editing, S.G.; visualization, Q.W.; supervision, S.G.; project administration, S.G.; funding acquisition, Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China grant number 41904117, Beijing Education Committee society branch plan surface project grant number KM202111232012, and Beijing Science and Technology Innovation Service Capacity Building-Fundamental Research Fund grant number PXM2019_014224_000026.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qiu, M.; Li, F.L.; Wang, S.; Gao, X.; Chen, Y.; Zhao, W.; Chen, H.; Huang, J.; Chu, W. AliMe Chat: A Sequence to Sequence and Rerank based Chatbot Engine. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 498–503. [Google Scholar]
  2. Zhou, L.; Gao, J.; Li, D.; Shum, H.Y. The Design and Implementation of XiaoIce, an Empathetic Social Chatbot. Comput. Linguist. 2020, 46, 53–93. [Google Scholar] [CrossRef]
  3. Haffner, P.; Tur, G.; Wright, J.H. Optimizing SVMs for complex call classification. In Proceedings of the 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, (ICASSP ‘03), Hong Kong, China, 6–10 April 2003. [Google Scholar]
  4. Collobert, R.; Weston, J. A unified architecture for natural language processing: Deep neural networks with multitask learning. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 160–167. [Google Scholar]
  5. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar]
  6. Lai, S.; Xu, L.; Liu, K.; Zhao, J. Recurrent convolutional neural networks for text classification. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  7. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, Z.H.; Min, F.; Chen, G.S.; Shen, S.P.; Wen, Z.C.; Zhou, X.B. Tri-Partition State Alphabet-Based Sequential Pattern for Multivariate Time Series. Cogn. Comput. 2021, 1–19. [Google Scholar] [CrossRef]
  9. Yang, Z.M.; Wang, L.Q. Research on application of deep learning algorithm in question intention classification. Comput. Eng. Appl. 2019, 55, 154–160. [Google Scholar]
  10. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  11. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  12. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  13. Joulin, E.; Grave, P.; Bojanowski, T. Mikolov, Bag of Tricks for Efficient Text Classification. arXiv 2016, arXiv:1607.01759. [Google Scholar]
  14. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. arXiv 2018, arXiv:1802.05365. [Google Scholar]
  15. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf (accessed on 10 November 2021).
  16. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  17. Li, S. Subword-level Chinese text classification method based on BERT. Comput. Sci. Appl. 2020, 10, 12677. [Google Scholar]
  18. Sun, Y.; Wang, S.; Li, Y.; Feng, S.; Chen, X.; Zhang, H.; Tian, X.; Zhu, D.; Tian, H.; Wu, H. ERNIE: Enhanced representation through knowledge integration. arXiv 2019, arXiv:1904.09223. [Google Scholar]
  19. Sau, B.B.; Balasubramanian, V.N. Deep model compression: Distilling knowledge from noisy teachers. arXiv 2016, arXiv:1610.09650. [Google Scholar]
  20. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  21. Liao, S.L.; Ji, J.M.; Yu, C.; Chen, X.P. Intent classification method based on BERT model and knowledge distillation. Comput. Eng. 2020, 10, 19678. [Google Scholar]
  22. Sun, S.; Cheng, Y.; Gan, Z.; Liu, J. Patient Knowledge Distillation for BERT Model Compression. arXiv 2019, arXiv:1908.09355. [Google Scholar]
Figure 1. Workflow of an e-commerce chatbot.
Figure 1. Workflow of an e-commerce chatbot.
Sensors 22 01270 g001
Figure 2. Structure of BERT model.
Figure 2. Structure of BERT model.
Sensors 22 01270 g002
Figure 3. Intention classification framework based on knowledge distillation.
Figure 3. Intention classification framework based on knowledge distillation.
Sensors 22 01270 g003
Figure 4. BERT and ERNIE mask mechanism.
Figure 4. BERT and ERNIE mask mechanism.
Sensors 22 01270 g004
Figure 5. FastText model mechanism.
Figure 5. FastText model mechanism.
Sensors 22 01270 g005
Figure 6. Accuracy of different algorithms.
Figure 6. Accuracy of different algorithms.
Sensors 22 01270 g006
Figure 7. Accuracy of Epoch with different dimensions of FastText model (-ngrams = 2; learning rate = 0.01; neg = 10).
Figure 7. Accuracy of Epoch with different dimensions of FastText model (-ngrams = 2; learning rate = 0.01; neg = 10).
Sensors 22 01270 g007
Figure 8. Accuracy at different -ngrams of FastText model (dimensions = 2; learning rate = 0.01; neg = 10).
Figure 8. Accuracy at different -ngrams of FastText model (dimensions = 2; learning rate = 0.01; neg = 10).
Sensors 22 01270 g008
Table 1. Part of the training corpus.
Table 1. Part of the training corpus.
CorpusClassification
I haven’t received the goods for so long.Urging Delivery
Why haven’t my shoes arrived for so long?Urging Delivery
Haven’t my things arrived yet?Urging Delivery
When will the exchanged goods come?Check the Replacement Progress
When will you send me a replacement?Check the Replacement Progress
When to exchange?Check the Replacement Progress
Where do I apply for exchange?Request a Replacement
I can’t find the place to apply for a replacement.Request a Replacement
The two pairs of children’s shoes I bought yesterday are small, how could I change them?Request A Replacement
Send it back by myself. What about the return shipping fee?Return Shipping Fee
How to calculate the return shipping fee for self-return?Return Shipping Fee
How to deal with the return shipping fee for self-return?Return Shipping Fee
Table 2. Super parameters of ERNIE and BERT model training.
Table 2. Super parameters of ERNIE and BERT model training.
ERNIE and BERT ParametersValue
Layer Numbers12
Learning Rate5 × 10−5
Batch Size256
Weight Decay0.01
Epoch10
Table 3. Parameters of FastText model training.
Table 3. Parameters of FastText model training.
FastText ParametersValue
Learning Rate0.01
neg10
ngram2
Dimension30
Table 4. Comparison of accuracy and time-consuming of different models.
Table 4. Comparison of accuracy and time-consuming of different models.
ModelAccuracy (%)Time-Consuming (ms)
Bi-LSTM94.3610.0+
FastText94.160.1
BERT (Base)94.5010.0+
ERNIE (Base)95.7010.0+
FastText (ERNIE Distillation)95.100.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, S.; Wang, Q. Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition. Sensors 2022, 22, 1270. https://doi.org/10.3390/s22031270

AMA Style

Guo S, Wang Q. Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition. Sensors. 2022; 22(3):1270. https://doi.org/10.3390/s22031270

Chicago/Turabian Style

Guo, Shiguang, and Qing Wang. 2022. "Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition" Sensors 22, no. 3: 1270. https://doi.org/10.3390/s22031270

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop