Adaption BERT for Medical Information Processing with ChatGPT and Contrastive Learning
Abstract
:1. Introduction
2. Related Work
3. Methods
3.1. Preliminary Knowledge
3.2. Overview
Algorithm 1 Similarity score s calculation process |
Input: a medical term , a medical term , Output: Similarity score s between and |
3.3. Pseudo-Label Generation with ChatGPT
3.4. Adaption BERT with Contrastive Loss
4. Experiments
4.1. Experiment Setup
4.2. Method Comparison
4.3. Qualitative Analysis
4.4. Ablation Studies
4.5. Limitations
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
Name | Meaning | |
Abbreviations | BERT | Bidirectional Encoder Representations from Transformers |
ChatGPT | Chat Generative Pre-trained Transformer | |
OpenEHR | open Electronic Health Record | |
MRR | Mean Reciprocal Rank | |
AUC | Area Under the Receiver Operating Characteristic Curve | |
MCC | Matthews Correlation Coefficient | |
EHR | Electronic Health Records | |
ICD-9-CM | International Classification of Diseases, 9th Revision, Clinical Modification | |
ICD-10-CM | International Classification of Diseases, 9th Revision, Clinical Modification | |
DDSSs | Digital Decision Support Systems | |
e-health | electronic health | |
Notations | OpenEHR-S | Pseudo-labeled OpenEHR dataset |
BERT-S | The model for processing the pseudo-labeled dataset. | |
E | World Embedding | |
Q | Query | |
K | Key | |
V | Value | |
T | The final layer vectors obtained after BERT processing | |
s | Similarity score | |
Ground Truth | ||
O | The average sum of vectors from the final layer obtained after BERT processing. | |
The margin of the contrastive loss function |
References
- Min, L.; Atalag, K.; Tian, Q.; Chen, Y.; Lu, X. Verifying the feasibility of implementing semantic interoperability in different countries based on the openEHR approach: Comparative study of acute coronary syndrome registries. JMIR Med. Inform. 2021, 9, e31288. [Google Scholar] [CrossRef] [PubMed]
- Kryszyn, J.; Cywoniuk, K.; Smolik, W.T.; Wanta, D.; Wróblewski, P.; Midura, M. Performance of an openEHR based hospital information system. Int. J. Med. Inform. 2022, 162, 104757. [Google Scholar] [CrossRef] [PubMed]
- Ferreira, D.E.; de Souza, J.M. Methodology for developing OpenEHR archetypes: A narrative literature review. J. Health Inform. 2023, 15, 53–59. [Google Scholar] [CrossRef]
- Talebi, S.; Tong, E.; Li, A.; Yamin, G.; Zaharchuk, G.; Mofrad, M.R. Exploring the performance and explainability of fine-tuned BERT models for neuroradiology protocol assignment. BMC Med. Inform. Decis. Mak. 2024, 24, 40. [Google Scholar] [CrossRef] [PubMed]
- Min, L.; Tian, Q.; Lu, X.; Duan, H. Modeling EHR with the openEHR approach: An exploratory study in China. BMC Med. Inform. Decis. Mak. 2018, 18, 1–15. [Google Scholar] [CrossRef] [PubMed]
- Min, L.; Tian, Q.; Lu, X.; An, J.; Duan, H. An openEHR based approach to improve the semantic interoperability of clinical data registry. BMC Med. Inform. Decis. Mak. 2018, 18, 49–56. [Google Scholar] [CrossRef] [PubMed]
- Johnson, D.; Goodman, R.; Patrinely, J.; Stone, C.; Zimmerman, E.; Donald, R.; Chang, S.; Berkowitz, S.; Finn, A.; Jahangir, E.; et al. Assessing the accuracy and reliability of AI-generated medical responses: An evaluation of the Chat-GPT model. Res. Sq. 2023, 28, rs.3.rs-2566942. [Google Scholar]
- Murali, L.; Gopakumar, G.; Viswanathan, D.M.; Nedungadi, P. Towards electronic health record-based medical knowledge graph construction, completion, and applications: A literature study. J. Biomed. Inform. 2023, 143, 104403. [Google Scholar] [CrossRef] [PubMed]
- Bertl, M.; Ross, P.; Draheim, D. Systematic AI support for decision-making in the healthcare sector: Obstacles and success factors. Health Policy Technol. 2023, 12, 100748. [Google Scholar] [CrossRef]
- Rossander, A.; Karlsson, D. Structure of Health Information With Different Information Models: Evaluation Study with Competency Questions. JMIR Med. Inform. 2023, 11, e46477. [Google Scholar] [CrossRef] [PubMed]
- Pedrera-Jiménez, M.; Kalra, D.; Beale, T.; Muñoz-Carrero, A.; Serrano-Balazote, P. Can OpenEHR, ISO 13606 and HL7 FHIR work together? An agnostic perspective for the selection and application of EHR standards from Spain. Authorea Prepr. 2023, 12, 100748. [Google Scholar]
- openEHR Website. Available online: https://www.openehr.org/ (accessed on 19 April 2024).
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Mutinda, F.W.; Yada, S.; Wakamiya, S.; Aramaki, E. Semantic textual similarity in Japanese clinical domain texts using BERT. Methods Inf. Med. 2021, 60, e56–e64. [Google Scholar] [CrossRef] [PubMed]
- Reese, J.T.; Danis, D.; Caufield, J.H.; Groza, T.; Casiraghi, E.; Valentini, G.; Mungall, C.J.; Robinson, P.N. On the limitations of large language models in clinical diagnosis. medRxiv 2023. [Google Scholar] [CrossRef] [PubMed]
- Vachatimanont, S.; Kingpetch, K. Exploring the capabilities and limitations of large language models in nuclear medicine knowledge with primary focus on GPT-3.5, GPT-4 and Google Bard. J. Med. Artif. Intell. 2024, 7. [Google Scholar] [CrossRef]
- Chakraborty, C.; Bhattacharya, M.; Lee, S.S. Need an AI-enabled, next-generation, advanced ChatGPT or large language models (LLMs) for error-free and accurate medical information. Ann. Biomed. Eng. 2024, 52, 134–135. [Google Scholar] [CrossRef] [PubMed]
- Walker, H.L.; Ghani, S.; Kuemmerli, C.; Nebiker, C.A.; Müller, B.P.; Raptis, D.A.; Staubli, S.M. Reliability of medical information provided by ChatGPT: Assessment against clinical guidelines and patient information quality instrument. J. Med. Internet Res. 2023, 25, e47479. [Google Scholar] [CrossRef] [PubMed]
- Cox, A.; Seth, I.; Xie, Y.; Hunter-Smith, D.J.; Rozen, W.M. Utilizing ChatGPT-4 for providing medical information on blepharoplasties to patients. Aesthetic Surg. J. 2023, 43, NP658–NP662. [Google Scholar] [CrossRef] [PubMed]
- Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Huggingface’s transformers: State-of-the-art natural language processing. arXiv 2019, arXiv:1910.03771. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
- Wang, J.; Huang, J.X.; Tu, X.; Wang, J.; Huang, A.J.; Laskar, M.T.R.; Bhuiyan, A. Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges. ACM Comput. Surv. 2024, 56, 1–33. [Google Scholar] [CrossRef]
- Rasmy, L.; Xiang, Y.; Xie, Z.; Tao, C.; Zhi, D. Med-BERT: Pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction. NPJ Digit. Med. 2021, 4, 86. [Google Scholar] [CrossRef] [PubMed]
- Liu, N.; Hu, Q.; Xu, H.; Xu, X.; Chen, M. Med-BERT: A pretraining framework for medical records named entity recognition. IEEE Trans. Ind. Inform. 2021, 18, 5600–5608. [Google Scholar] [CrossRef]
- Luo, L.; Ning, J.; Zhao, Y.; Wang, Z.; Ding, Z.; Chen, P.; Fu, W.; Han, Q.; Xu, G.; Qiu, Y.; et al. Taiyi: A bilingual fine-tuned large language model for diverse biomedical tasks. J. Am. Med. Inform. Assoc. 2024, ocae037. [Google Scholar] [CrossRef] [PubMed]
- He, J.; Li, P.; Liu, G.; Zhao, Z.; Zhong, S. PeFoMed: Parameter Efficient Fine-tuning on Multimodal Large Language Models for Medical Visual Question Answering. arXiv 2024, arXiv:2401.02797. [Google Scholar]
- Shi, W.; Xu, R.; Zhuang, Y.; Yu, Y.; Wu, H.; Yang, C.; Wang, M.D. MedAdapter: Efficient Test-Time Adaptation of Large Language Models towards Medical Reasoning. arXiv 2024, arXiv:2405.03000. [Google Scholar]
- Muizelaar, H.; Haas, M.; van Dortmont, K.; van der Putten, P.; Spruit, M. Extracting Patient Lifestyle Characteristics from Dutch Clinical Text with BERT Models. BMC Med. Inform. Decis. Mak. 2024, 24, 151. [Google Scholar] [CrossRef] [PubMed]
- Kulkarni, D.; Ghosh, A.; Girdhari, A.; Liu, S.; Vance, L.A.; Unruh, M.; Sarkar, J. Enhancing pre-trained contextual embeddings with triplet loss as an effective fine-tuning method for extracting clinical features from electronic health record derived mental health clinical notes. Nat. Lang. Process. J. 2024, 6, 100045. [Google Scholar] [CrossRef]
- Kumar, P.S. Bridging the Knowledge Gap: Improving BERT models for answering MCQs by using Ontology-generated synthetic MCQA Dataset. In Proceedings of the International FLAIRS Conference Proceedings, Sandestin Beach, FL, USA, 19–21 May 2024; Volume 37. [Google Scholar]
- Suneetha, A.R.N.; Mahalngam, T. Fine tuning bert based approach for cardiovascular disease diagnosis. Int. J. Intell. Syst. Appl. Eng. 2023, 11, 59–66. [Google Scholar]
- Kim, Y.; Kim, J.H.; Kim, Y.M.; Song, S.; Joo, H.J. Predicting medical specialty from text based on a domain-specific pre-trained BERT. Int. J. Med. Inform. 2023, 170, 104956. [Google Scholar] [CrossRef] [PubMed]
- Su, P.; Vijay-Shanker, K. Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction. BMC Bioinform. 2022, 23, 120. [Google Scholar] [CrossRef] [PubMed]
- Ding, J.; Li, B.; Xu, C.; Qiao, Y.; Zhang, L. Diagnosing crop diseases based on domain-adaptive pre-training BERT of electronic medical records. Appl. Intell. 2023, 53, 15979–15992. [Google Scholar] [CrossRef]
- Babu, A.; Boddu, S.B. BERT-Based Medical Chatbot: Enhancing Healthcare Communication through Natural Language Understanding. Explor. Res. Clin. Soc. Pharm. 2024, 13, 100419. [Google Scholar] [CrossRef] [PubMed]
- Chen, H.; Qin, D.; Zhang, X.; Zhang, H.; Liang, X. A Study on the Classification of Chinese Medicine Records Using BERT, Chest Impediment as an Example. In Proceedings of the CCF International Conference on Natural Language Processing and Chinese Computing, Guilin, China, 24–25 September 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 29–37. [Google Scholar]
- Faris, H.; Faris, M.; Habib, M.; Alomari, A. Automatic symptoms identification from a massive volume of unstructured medical consultations using deep neural and BERT models. Heliyon 2022, 8, e09683. [Google Scholar] [CrossRef] [PubMed]
- Zheng, T.; Gao, Y.; Wang, F.; Fan, C.; Fu, X.; Li, M.; Zhang, Y.; Zhang, S.; Ma, H. Detection of medical text semantic similarity based on convolutional neural network. BMC Med. Inform. Decis. Mak. 2019, 19, 1–11. [Google Scholar] [CrossRef] [PubMed]
- Liang, H.; Lin, K.; Zhu, S. Short text similarity hybrid algorithm for a Chinese medical intelligent question answering system. In Proceedings of the Technology-Inspired Smart Learning for Future Education: 29th National Conference on Computer Science Technology and Education, NCCSTE 2019, Kaifeng, China, 9–11 October 2019; Revised Selected Papers 29. Springer: Berlin/Heidelberg, Germany, 2020; pp. 129–142. [Google Scholar]
- Li, Q.; He, S. Similarity matching of medical question based on Siamese network. BMC Med. Inform. Decis. Mak. 2023, 23, 55. [Google Scholar] [CrossRef] [PubMed]
- Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S.; et al. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://www.mikecaptain.com/resources/pdf/GPT-1.pdf (accessed on 20 May 2024).
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. Openai Blog 2019, 1, 9. [Google Scholar]
- Dey, R.K.; Das, A.K. Modified term frequency-inverse document frequency based deep hybrid framework for sentiment analysis. Multimed. Tools Appl. 2023, 82, 32967–32990. [Google Scholar] [CrossRef]
- Wan, Q.; Xu, X.; Han, J. A dimensionality reduction method for large-scale group decision-making using TF-IDF feature similarity and information loss entropy. Appl. Soft Comput. 2024, 150, 111039. [Google Scholar] [CrossRef]
- Selva Birunda, S.; Kanniga Devi, R. A review on word embedding techniques for text classification. In Innovative Data Communication 436 Technologies and Application: Proceedings of ICIDCA 2020; Springer: Berlin/Heidelberg, Germany, 2021; Volume 59, pp. 267–281. [Google Scholar]
- Berger, B.; Waterman, M.S.; Yu, Y.W. Levenshtein distance, sequence comparison and biological database search. IEEE Trans. Inf. Theory 2020, 67, 3287–3294. [Google Scholar] [CrossRef] [PubMed]
- Zhao, C.; Sahni, S. String correction using the Damerau–Levenshtein distance. BMC Bioinform. 2019, 20, 1–28. [Google Scholar] [CrossRef] [PubMed]
- ICD-9-CM Website. Available online: https://www.cms.gov/medicare/coding-billing/icd-10-codes/icd-9-cm-diagnosis-procedure-codes-abbreviated-and-full-code-titles/ (accessed on 10 June 2024).
- ICD-10-CM Website. Available online: https://www.cms.gov/medicare/coding-billing/icd-10-codes/2024-icd-10-cm/ (accessed on 10 June 2024).
Methods | Authors | Problem | Solution |
---|---|---|---|
CNN-based | Zheng et al. [38] | Manually retrieving and comparing imaging and pathology reports with overlapping exam body sites is time-consuming. | A convolutional neural network model was used to calculate similarities among report pairs. |
Liang et al. [39] | Calculating the semantic similarity of noisy short medical question texts for an intelligent QA system. | A shared layer-based CNN combined with TF-IDF for feature extraction and noise reduction. | |
Li et al. [40] | Enhancing the efficiency of online medical QA by accurately matching user questions with professional medical answers. | A bidirectional gated recurrent unit network with CNN for feature extraction to improve matching accuracy. | |
BERT-Based | Suneetha et al. [31] | Improve the diagnosis effect of cardiovascular disease. | Fine-tuning BERT to provide an effective and accurate diagnosis. |
Kim et al. [32] | Use natural language processing technology to improve outpatient diagnosis and treatment initiation efficiency and accuracy. | A medical professional prediction model based on BERT patient-side medical question text is used to diagnose outpatients. | |
Su et al. [33] | Improve the accuracy of automatic extraction of biomedical relationships. | Pre-training and fine-tuning BERT to improve extraction accuracy. | |
Ding et al. [34] | Crop disease diagnosis. | BERT and RCNN-based model to assist plant doctors in diagnosing crop diseases. | |
Babu et al. [35] | Healthcare communication and accessibility. | BERT-based medical chatbot using advanced deep-learning techniques. | |
Chen et al. [36] | Text classification for coronary heart disease. | BERT-based pre-trained diagnostic model for traditional Chinese medicine texts. | |
Faris et al. [37] | Symptom identification and diagnosis in multiple languages. | BERT-based method to assist doctors in handling multilingual consultations. |
Target Text | Pseudo-Label |
---|---|
exclusion of pregnancy | Negative pregnancy test |
iciq-ui short form | Urinary Incontinence Questionnaire - Short Version |
financial record | Financial documentation |
visual acuity test result | Vision clarity assessment outcome |
palpation of the uterus | Manual examination of the uterus |
organisation | institution |
breast-feeding summary | Lactation overview |
oocyte and embryo assessment | Examination of oocyte and embryo development |
examination of the cornea | Corneal examination |
informed consent | Informed agreement |
clinical synopsis | Clinical summary |
housing summary | Living situation summary |
inspection of the vagina | Vaginal examination |
self-test result | Personal assessment outcome |
classification of glaucoma | Glaucoma categorization |
imaging examination of a pregnant uterus | Antenatal imaging of the uterus |
inspection of the rectum | Rectal examination |
imaging examination of a placenta | Placental imaging study |
occupation record | Career profile |
imaging examination of the scrotum | Scrotal imaging |
Models | Top 1 | MRR | AUC | MCC | |||
---|---|---|---|---|---|---|---|
TF-IDF | 0.0 | 22.8 | 75.0 | 50.0 | 60.0 | 67.5 | 13.2 |
Count Vector | 25.0 | 77.0 | 75.0 | 50.0 | 60.0 | 66.7 | 13.2 |
Levenshtein Distance | 35.0 | 85.8 | 75.5 | 55.0 | 63.6 | 76.1 | 15.9 |
Damerau–Levenshtein Distance | 35.0 | 85.8 | 75.5 | 55.0 | 63.6 | 73.4 | 14.7 |
BERT | 85.0 | 86.7 | 79.0 | 90.0 | 84.1 | 95.3 | 34.4 |
Ours | 85.0 | 87.3 | 79.5 | 95.0 | 86.6 | 95.5 | 37.1 |
Models | Top 1 | MRR | AUC | MCC | ||||
---|---|---|---|---|---|---|---|---|
Group1 | TF-IDF | 0.0 | 10.8 | 97.0 | 20.0 | 33.2 | 66.1 | 7.9 |
Count Vector | 5.0 | 15.8 | 96.8 | 15.0 | 29.8 | 66.3 | 5.6 | |
Levenshtein Distance | 10.0 | 37.0 | 97.2 | 35.0 | 51.5 | 73.5 | 14.7 | |
Damerau–Levenshtein Distance | 5.0 | 32.0 | 97.2 | 35.0 | 51.5 | 73.4 | 14.7 | |
BERT | 25.0 | 35.6 | 97.4 | 55.0 | 70.3 | 94.1 | 23.8 | |
Ours | 30.0 | 40.2 | 97.4 | 60.0 | 74.3 | 94.7 | 26.1 | |
Group2 | TF-IDF | 10.0 | 20.0 | 97.1 | 25.0 | 40.0 | 66.6 | 10.4 |
Count Vector | 10.0 | 22.0 | 97.0 | 20.0 | 33.2 | 66.5 | 7.9 | |
Levenshtein Distance | 15.0 | 37.8 | 97.1 | 30.0 | 45.8 | 76.6 | 12.4 | |
Damerau–Levenshtein Distance | 15.0 | 37.8 | 97.1 | 30.0 | 45.8 | 73.4 | 14.7 | |
BERT | 35.0 | 47.2 | 97.5 | 70.0 | 81.5 | 94.8 | 30.6 | |
Ours | 50.0 | 56.7 | 97.5 | 70.0 | 81.5 | 95.4 | 30.6 | |
Group3 | TF-IDF | 5.0 | 17.5 | 97.0 | 20.0 | 33.2 | 66.1 | 7.0 |
Count Vector | 10.0 | 20.0 | 97.0 | 20.0 | 33.2 | 66.2 | 7.9 | |
Levenshtein Distance | 15.0 | 40.0 | 97.1 | 25.0 | 40.0 | 72.7 | 10.1 | |
Damerau–Levenshtein Distance | 15.0 | 40.0 | 97.1 | 25.0 | 40.0 | 73.4 | 14.7 | |
BERT | 35.0 | 46.8 | 97.6 | 70.0 | 81.5 | 95.1 | 32.9 | |
Ours | 40.0 | 47.9 | 97.4 | 60.0 | 74.3 | 95.1 | 26.1 | |
Group4 | TF-IDF | 0.0 | 9.0 | 97.0 | 20.0 | 33.2 | 65.2 | 7.9 |
Count Vector | 5.0 | 14.5 | 97.1 | 15.0 | 40.0 | 65.4 | 10.1 | |
Levenshtein Distance | 10.0 | 33.3 | 97.2 | 35.0 | 51.5 | 73.4 | 14.7 | |
Damerau–Levenshtein Distance | 10.0 | 33.3 | 97.2 | 35.0 | 51.5 | 73.4 | 14.7 | |
BERT | 30.0 | 39.9 | 97.4 | 60.0 | 74.3 | 94.7 | 26.1 | |
Ours | 30.0 | 40.2 | 97.4 | 60.0 | 74.3 | 94.9 | 26.1 |
Dataset | Models | Top 1 | MRR | AUC | MCC | FLOPs | Parameters | |||
---|---|---|---|---|---|---|---|---|---|---|
ICD-10-CM | BERT | 80.0 | 86.7 | 83.7 | 96.0 | 89.4 | 98.0 | 38.8 | 1.19G | 104.4M |
Ours | 80.0 | 87.3 | 83.7 | 96.0 | 89.4 | 98.2 | 38.8 | 1.19G | 104.4M | |
ICD-9-CM | BERT | 40.0 | 46.1 | 81.1 | 64.0 | 68.9 | 79.2 | 22.5 | 1.19G | 104.4M |
Ours | 36.0 | 49.9 | 81.8 | 72.0 | 76.6 | 82.5 | 26.5 | 1.19G | 104.4M |
Target Text | Pseudo Label | BERT Prediction | Our Prediction |
---|---|---|---|
self-test result | Personal assessment outcome | Personal assessment outcome Test sample Pregnancy test result Specimen measurements Examination findings-lens | Personal assessment outcome Health assessment questionnaire Neonatal assessment score Exposure screening questionnaire Acquisition details on visual field test |
classification of glaucoma | Glaucoma categorization | Glaucoma categorization Examination of a tympanic membrane Examination of the respiratory system Fundoscopic examination of eyes Scrotal imaging study | Glaucoma categorization Examination of the respiratory system Jugular venous pressure Scrotal imaging study Fundoscopic examination of eyes |
inspection of the rectum | Rectal examination | Examination of the thyroid Examination of a breast Palpation of the prostate Examination of the respiratory system Rectal examination | Examination of the thyroid Examination of a breast Rectal examination Examination of the respiratory system Palpation of the prostate |
imaging examination of a placenta | Placental imaging study | Imaging examination of a body structure Placental imaging study Obstetric ultrasound scan Scrotal imaging study Ophthalmic tomography examination | Placental imaging study Imaging examination of a body structure Scrotal imaging study Obstetric ultrasound scan Ophthalmic tomography examination |
Min Positive | Max Negative | Top 1 | MRR | AUC | MCC | ||||
---|---|---|---|---|---|---|---|---|---|
Group1 | 25.0 | 35.6 | 97.4 | 55.0 | 70.3 | 94.1 | 23.8 | ||
✓ | 30.0 | 39.2 | 97.4 | 55.0 | 70.3 | 94.6 | 23.8 | ||
✓ | 30.0 | 39.3 | 97.4 | 55.0 | 70.3 | 94.5 | 23.8 | ||
✓ | ✓ | 30.0 | 40.2 | 97.4 | 60.0 | 74.3 | 94.7 | 26.1 | |
Group2 | 35.0 | 47.2 | 97.5 | 70.0 | 81.5 | 94.8 | 30.6 | ||
✓ | 35.0 | 49.8 | 97.6 | 75.0 | 84.8 | 95.4 | 32.9 | ||
✓ | 50.0 | 53.7 | 97.5 | 70.0 | 81.5 | 95.1 | 30.6 | ||
✓ | ✓ | 50.0 | 56.7 | 97.5 | 70.0 | 81.5 | 95.4 | 30.6 | |
Group3 | 35.0 | 46.8 | 97.6 | 70.0 | 81.5 | 95.1 | 32.9 | ||
✓ | 35.0 | 45.6 | 97.5 | 65.0 | 78.0 | 95.3 | 28.3 | ||
✓ | 35.0 | 45.6 | 97.4 | 60.0 | 74.3 | 94.8 | 26.1 | ||
✓ | ✓ | 40.0 | 47.9 | 97.4 | 60.0 | 74.3 | 95.1 | 26.1 | |
Group4 | 30.0 | 39.9 | 97.4 | 60.0 | 74.3 | 94.7 | 26.1 | ||
✓ | 25.0 | 35.6 | 97.5 | 65.0 | 70.3 | 94.9 | 26.1 | ||
✓ | 30.0 | 39.9 | 97.4 | 60.0 | 74.3 | 94.9 | 26.1 | ||
✓ | ✓ | 30.0 | 40.2 | 97.4 | 60.0 | 74.3 | 94.9 | 26.1 |
Models | Top 1 | MRR | AUC | MCC | |||
---|---|---|---|---|---|---|---|
Baseline | 85.0 | 86.7 | 79.0 | 90.0 | 84.1 | 95.3 | 34.4 |
Cross Entropy Loss | 80.0 | 83.8 | 79.0 | 90.0 | 84.1 | 94.4 | 34.4 |
Smooth Cross Entropy Loss | 75.0 | 81.3 | 79.0 | 90.0 | 84.1 | 94.2 | 34.4 |
Contrastive Loss* | 85.0 | 86.7 | 79.0 | 90.0 | 84.1 | 95.3 | 34.4 |
Contrastive Loss | 85.0 | 87.3 | 79.5 | 95.0 | 86.6 | 95.5 | 37.1 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Min, L.; Fan, Z.; Dou, F.; Sun, J.; Luo, C.; Lv, Q. Adaption BERT for Medical Information Processing with ChatGPT and Contrastive Learning. Electronics 2024, 13, 2431. https://doi.org/10.3390/electronics13132431
Min L, Fan Z, Dou F, Sun J, Luo C, Lv Q. Adaption BERT for Medical Information Processing with ChatGPT and Contrastive Learning. Electronics. 2024; 13(13):2431. https://doi.org/10.3390/electronics13132431
Chicago/Turabian StyleMin, Lingtong, Ziman Fan, Feiyang Dou, Jiaao Sun, Changsheng Luo, and Qinyi Lv. 2024. "Adaption BERT for Medical Information Processing with ChatGPT and Contrastive Learning" Electronics 13, no. 13: 2431. https://doi.org/10.3390/electronics13132431
APA StyleMin, L., Fan, Z., Dou, F., Sun, J., Luo, C., & Lv, Q. (2024). Adaption BERT for Medical Information Processing with ChatGPT and Contrastive Learning. Electronics, 13(13), 2431. https://doi.org/10.3390/electronics13132431