Next Article in Journal
Kyoto Classification-Based Predictive Factors Associated with the Development of Gastric Cancer After Helicobacter pylori Eradication: A Prospective Multicenter Observational Study
Previous Article in Journal
Heterotopic Cesarean Scar Pregnancy: A Systematic Review of Diagnosis, Management and Prognosis
Previous Article in Special Issue
The Prediction of Radiation-Induced Trismus by the Apparent Diffusion Coefficient Values of Masseter Muscles before Chemoradiotherapy in Locally Advanced Nasopharyngeal Carcinomas
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Confirmation of Large Language Models in Head and Neck Cancer Staging

by
Mehmet Kayaalp
1,2,
Hatice Bölek
1,2 and
Hatime Arzu Yaşar
1,2,*
1
Department of Medical Oncology, Ankara University Faculty of Medicine, Ankara University, Ankara 06620, Turkey
2
Ankara University Cancer Research Institute, Ankara University, Ankara 06620, Turkey
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(18), 2375; https://doi.org/10.3390/diagnostics15182375
Submission received: 10 August 2025 / Revised: 10 September 2025 / Accepted: 13 September 2025 / Published: 18 September 2025
(This article belongs to the Special Issue Integrative Approaches in Head and Neck Cancer Imaging)

Abstract

Background/Objectives: Head and neck cancer (HNC) is a heterogeneous group of malignancies in which staging plays a critical role in guiding treatment and prognosis. Large language models (LLMs) such as ChatGPT, DeepSeek, and Grok have emerged as potential tools in oncology, yet their clinical applicability in staging remains unclear. This study aimed to evaluate the accuracy and concordance of LLMs compared to clinician-assigned staging in patients with HNC. Methods: The medical records of 202 patients with HNC, who presented to our center between 1 January 2010 and 13 February 2025, were retrospectively reviewed. The information obtained from the hospital information system by a junior researcher was re-evaluated by a senior researcher, and standard staging was completed. Except for the stage itself, the data used for staging were provided to a blinded third researcher, who then entered them into the ChatGPT, DeepSeek, and Grok applications with a staging command. After all staging processes were completed, the data were compiled, and clinician-assigned stages were compared with those generated by the LLMs. Results: The majority of the patients had laryngeal (45.5%) and nasopharyngeal cancer (21.3%). Definitive surgery was performed in 39.6% of the patients. Stage 4 was the most common stage among the patients (54%). The overall concordance rates, Cohen’s kappa values, and F1 scores were 85.6%, 0.797, and 0.84 for ChatGPT; 67.3%, 0.522, and 0.65 for DeepSeek; and 75.2%, 0.614, and 0.72 for Grok, respectively, with no statistically significant differences between models. Pathological and surgical staging were found to be similar in terms of concordance. The concordance of assessments utilizing only imaging, only pathology notes, only physical examination notes, and comprehensive information was evaluated, revealing no significant differences. Conclusions: Large language models (LLMs) demonstrate relatively high accuracy in staging HNC. With careful implementation and with the consideration of prospective studies, these models have the potential to become valuable tools in oncology practice.

1. Introduction

Head and neck cancers (HNCs) represent a significant global health issue, with approximately 946,456 new cases and 481,001 cancer-related deaths reported in the 2022 GLOBOCAN data [1]. The most common subtypes of HNCs include cancers of the lip and oral cavity, larynx, nasopharynx, oropharynx, and hypopharynx. The etiology of head and neck cancers includes environmental factors such as tobacco and alcohol consumption, viral agents including human papillomavirus (HPV) and Epstein–Barr virus (EBV), and genetic syndromes. The treatment of head and neck cancers includes surgical resection, definitive chemoradiotherapy, and trimodal therapy, with the TNM staging system serving as a key prognostic factor in determining the optimal therapeutic approach [2].
The oral cavity is bounded anteriorly by the lip mucosa and posteriorly by the anterior tonsillar region. The nasopharynx extends from the clivus superiorly to the hard and soft palate inferiorly. The oropharynx ranges from the circumvallate papillae to the posterior pharyngeal wall. The hypopharynx includes the piriform sinuses and the postcricoid area. The larynx extends from the supraglottic region to below the cricoid cartilage. The paranasal sinuses include the maxillary, ethmoidal, sphenoid, and frontal regions. The salivary glands consist of the parotid, submandibular, and sublingual glands.
In the staging algorithm, following a physical examination and endoscopic assessment conducted by an otolaryngologist specialized in oncology, imaging modalities such as computed tomography (CT), magnetic resonance imaging (MRI), and positron emission tomography (PET), along with histopathological evaluation via biopsy, are essential for comprehensive assessment [3]. The most common histological variant is squamous cell carcinoma; however, adenocarcinoma in glandular tumors, as well as melanoma and sarcoma, may also be observed. Testing for EBV and HPV, which are potential etiological factors for squamous cell carcinoma, is essential for staging, prognosis, and follow-up.
In the staging of head and neck cancer, the American Joint Committee on Cancer (AJCC) TNM 8 staging system is utilized, which may vary based on clinical, pathological, and molecular characteristics [4]. Preoperative clinical staging is performed not only based on tumor size but also according to anatomical landmarks. The necessity of evaluation by a specialized radiologist and the omission of critical anatomical landmarks in imaging reports complicate the staging process. Separate staging systems are utilized for laryngeal cancer, including the supraglottic, glottic, and subglottic regions, as well as for sinus cancers, which encompass the maxillary sinus, ethmoid sinus, and nasal cavity. In the staging of these malignancies, along with nasopharyngeal cancer, anatomical localization is prioritized over tumor size. Additionally, oropharyngeal cancer staging incorporates two distinct systems based on p16 positivity. Given that both clinical and pathological staging are applied to all head and neck cancers except nasopharyngeal cancer, the staging process is considered complex. These factors raise an important question about the feasibility of employing artificial intelligence applications as tools in the staging process.
Large language models (LLMs) have recently transformed the landscape of artificial intelligence, demonstrating exceptional capabilities in natural language processing (NLP) tasks [5]. These advances stem from the evolution of neural architectures, particularly transformers, alongside increased computational power and expansive training datasets. LLMs can perform complex language-related tasks such as translation, summarization, information retrieval, and conversational interactions with human-like proficiency. Natural language processing (NLP) models, through their integration with fields such as proteomics, are increasingly contributing to oncology by supporting tasks like tumor T-cell antigen (TTCA) prediction, which is critical for the development of personalized cancer therapies [6,7]. The scalability of these models—ranging from billions to hundreds of billions of parameters—has significantly improved their performance and generalization capabilities.
The evolution of large language models began in 2019 with T5 and mT5. In 2020, OpenAI introduced GPT-3, marking a major advancement in AI. In 2021, models like Codex, Jurassic-1, LaMDA, and GLM were developed. By 2022, LLaMA, BLOOM, PaLM, and OPT gained prominence, while 2023 saw the rise of LLaMA 2, Falcon, Mistral, and StarCoder. By 2024, models like GPT-4o, Gemini 1.5, Claude, Mixtral 8x22B, and DeepSeek V2 represented the latest advancements in artificial intelligence [5]. Launched in January 2025, the China-based DeepSeek-R1, developed with cost-efficient methods and limited computational resources, garnered significant attention for exhibiting performance comparable to that of ChatGPT o1 [8]. Grok is an LLM-based chatbot developed by xAI, with access to X (formerly Twitter).
Numerous studies have been conducted to validate LLM models, particularly in the fields of oncology and medicine. Although these models aim to improve healthcare accessibility, they may yield inaccurate outcomes that can result in medical errors [9].
Most oncology-related research focuses on confirming established medical knowledge, while studies evaluating patient data remain relatively scarce [10].
As pre-trained language models have been tested in cancer staging [11], studies on the use of widely accessible and commonly used LLM models for staging are rapidly emerging in the literature [11]. Staging studies conducted with ChatGPT, which is considered the most commonly used chatbot in head and neck cancers in general and oropharyngeal cancers in particular, are available [12,13]. However, no such study has yet been conducted for DeepSeek or Grok.
This study was designed to explore the use of ChatGPT while also evaluating DeepSeek and Grok for the first time in both oncology and head and neck cancer staging, aiming to assess their potential in clinical applications.

2. Materials and Methods

2.1. Patient Group

A total of 202 patients with head and neck cancer who presented to Ankara University medical oncology clinic between 1 January 2010 and 13 February 2025 were included in the study. Patient files were compiled by reviewing outpatient clinic notes to obtain clinical examination records (physical examination, endoscopic images), radiological imaging [ultrasonography (USG), CT, MRI, PET-CT, etc.], pathology results, and surgical notes. Patients with recurrence were excluded due to possible anatomical alterations following treatments.
The sample size of 202 patients in our study provides approximately 85% power to detect a 10% absolute difference in diagnostic accuracy (e.g., 76% vs. 86%) at a two-sided significance level of 0.05, based on estimates derived from the study by Tozuka et al., which used 100 fictional cases to evaluate large language model performance in lung cancer staging [14].

2.2. Staging by Clinicians

The compiled clinical notes were initially staged by a junior medical oncologist, and subsequently reviewed by a senior medical oncologist to establish a reference standard staging. Staging was performed according to the AJCC 8th edition TNM classification. Pathological staging was used for patients who underwent definitive surgery, whereas clinical staging was applied for patients who did not undergo definitive surgery.

2.3. LLM-Based Cancer Staging

After clinician-assigned staging was entered into the database, the patient files containing the information required for staging were forwarded to a third researcher who was blinded to the final staging results. This researcher submitted the anonymized data to ChatGPT, Grok, and DeepSeek, using only the prompt “What is the TNM stage?”. Supplementary Figure S1 illustrates an example of this process. To better simulate a real-world clinical scenario and to evaluate both accurate and hallucinated outputs, no prompt engineering was applied. Instead, a single standardized query was used across all cases, reflecting the way a clinician might naturally interact with such systems rather than optimizing prompts for maximal performance. Due to the absence of analysis limitations, ChatGPT Plus 4O [15] was used for ChatGPT, V3 for DeepSeek [16], and Grok3 for Grok [17]. All models were accessed via publicly available open web interfaces. Patient data intended for uploading to LLMs were anonymized by removing personal identifiers such as name, age, city, hospital information, and pathology numbers. Each patient was analyzed in a separate chat session to prevent potential self-training effects within a session.

2.4. Statistical Analysis

The staging results were entered into SPSS Version 26 [18], where the clinician-assessed staging data were also recorded, and the analysis was conducted. The chi-square test was used to assess the relationship between categorical variables, while Fisher’s exact test was applied for small sample groups. Cochran’s Q test was used to analyze three or more dependent categorical variables. Following a significant Cochran’s Q result, McNemar tests were performed for pairwise post hoc comparisons between models. Categorical agreement between large language models (LLMs) and the reference staging was assessed using Cohen’s kappa coefficient, which quantifies inter-rater agreement beyond chance. Kappa values were interpreted based on Landis and Koch’s criteria. For performance evaluation, macro-averaged F1 scores were calculated for each model by comparing their predicted T, N, M, and overall stage values with the corresponding ground truth labels on a case-by-case basis. All p-values were calculated based on a 95% confidence level. For all pairwise comparisons involving either McNemar tests or kappa coefficient comparisons, statistical significance was determined using the Bonferroni correction.

2.5. Ethical Approval and Informed Consent

Ethical approval was obtained from the Ankara University Faculty of Medicine Human Research Ethics Committee on 13 March 2025, with the approval number 2025/168. All researchers adhered to the principles of the Declaration of Helsinki and provided their signed consent. Given the retrospective design and use of fully anonymized data, the requirement for individual informed consent was waived by the Ethics Committee.
The workflow of the study is shown in Figure 1.

3. Results

3.1. Demographics and Patient Characteristics

The majority of the patients were male (74.8%). Regarding smoking history, 37.1% were active smokers, 36.1% were former smokers, and 25.2% had never smoked. The median age was 58.35 (SD: 14.35), with the most common disease localization being the larynx (45.5%). Only 39.6% of the patients underwent definitive surgery. Patient characteristics are shown in Table 1. For staging, complete data—including pathology reports, imaging, and clinical examination notes—were available for 149 patients (73.8%) (Table 1).
Among all patients, the most common tumor stage was T4 (34.6%), and the most frequently observed nodal stage was N0 (36.1%). All patients were classified as M0. Overall, stage IV was the most prevalent disease stage, observed in 54.4% of cases (Supplementary Table S1).

3.2. Staging Results by Clinicians and Large Language Models

In assessing concordance with clinicians, ChatGPT demonstrated the highest agreement with 173 matches (85.6%), followed by Grok with 152 matches (75.2%) and DeepSeek with 136 matches (67.3%) (Table 2). The agreement in the metastasis (M) category was the highest among all of them, with a concordance rate of 100%. Significant differences were observed between ChatGPT and both DeepSeek (p < 0.001) and Grok (p = 0.009), indicating the superior performance of ChatGPT in staging accuracy. No significant difference was found between DeepSeek and Grok (p = 0.077) (Table 2).
Cohen’s kappa analysis revealed that ChatGPT had the highest level of agreement. All large language models demonstrated statistically significant agreement across T, N, M, and stage values according to Cohen’s kappa analysis (Figure 2, Supplementary Table S2).
The comparison of agreement and kappa statistics revealed no statistically significant difference between clinically and surgically staged patients (Supplementary Table S3). No statistically significant difference was found between the agreement rates of evaluations based on imaging, pathology, physical examination findings, and the combination of all available data (Supplementary Table S4).
Across all tumor localizations, the kappa values indicated consistent agreement for each model (Table 3).
Macro-averaged F1 scores were calculated by comparing each model’s predicted T, N, M, and overall stage values with the ground truth labels on a per-patient basis. ChatGPT achieved an F1 score of 0.78 for T classification, with DeepSeek and Grok following at 0.69 and 0.64, respectively. In N classification, scores showed greater variability; ChatGPT reached 0.86, while DeepSeek and Grok recorded 0.66 and 0.78, respectively. Overall stage prediction resulted in F1 scores of 0.85 for ChatGPT, 0.65 for DeepSeek, and 0.72 for Grok. All models correctly predicted M0 in all cases, resulting in an F1 score of 1.0 for M (Figure 3, Supplementary Table S5).
Concordance rates stratified by gender, surgical status, and TNM stage are presented in Supplementary Tables S6–S8. While no significant differences were observed according to gender or surgical status, a statistically significant variation in concordance was noted across TNM stages.

4. Discussion

Artificial intelligence applications are increasingly being integrated into medicine and oncology [19]. LLMs represent one of the fastest evolving and most widely adopted areas of AI by the public. To our knowledge, this is the first study to evaluate Grok and DeepSeek—two newly developed LLMs—for clinical use in oncology, specifically in the staging of HNCs.
Despite the advancements in artificial intelligence and LLM technologies, common issues such as errors and hallucinations observed across all LLMs can lead to critical consequences when used in the field of healthcare [20]. Therefore, many validation studies are currently underway to assess the reliability of LLMs in medical settings. In addition to widely used general-purpose LLMs, models specifically developed for healthcare are also being evaluated in various studies. In oncology and head and neck cancer, conformity studies with LLMs primarily focus on answering patient questions that could be directed to clinicians, answering board exam questions, and determining treatment plans [21,22,23,24,25,26,27].
Staging serves as a key predictor in both oncologic prognosis and treatment planning. Therapeutic strategies recommended by international guidelines are guided not solely by tumor resectability, but primarily by the disease stage. For example, in T1N0 nasopharyngeal carcinoma, definitive radiotherapy alone is recommended [28]. From T2N0 onward, concurrent chemoradiotherapy (CRT) becomes the standard approach. While the NCCN guideline recommends evaluating T3 disease for induction chemotherapy, the ESMO guideline explicitly supports induction chemotherapy in patients with T3N1 disease. In cases with N2 involvement, ESMO recommends adjuvant chemotherapy following CRT [29]. Likewise, T4N0 disease is also considered an indication for induction chemotherapy according to ESMO. In laryngeal cancer, the ESMO guideline recommends laser cordectomy as a treatment option for T1 tumors, whereas concurrent CRT is preferred for T1–2 tumors with N2–3 nodal disease [2]. Similarly, for oral cavity cancers, the ESMO guideline does not recommend surgery for cT4b tumors, emphasizing the importance of stage-specific non-surgical approaches in such cases [2].
There are many studies in the literature on the use of artificial intelligence applications in cancer staging. In a study leveraging pathology reports for cancer staging, the highest F1 score for T classification was achieved by the fine-tuned Clinical BigBird model (F1 = 0.81). For N classification, domain-specific clinical large language models—ClinicalCamel-70B and Med42-70B—demonstrated superior performance (F1 = 0.82–0.81) compared to the general-purpose LLaMA-2-based LLM [30].
Machine learning and large language model-based approaches have been investigated for head and neck cancer staging. In a recent study focusing on oropharyngeal cancer, the accuracy of automated T and N staging remained relatively limited (T: 55.9%, N: 56.0%), whereas M staging achieved higher performance (87.6%), in contrast to the higher staging accuracy observed in our study. In patients with lung cancer, the staging accuracy of ChatGPT using radiology and pathology reports has been reported to range between 70% and 99%, depending on the input type and the specific component of the TNM staging system [31,32]. A summary of recent studies evaluating LLMs in oncology and cancer staging is provided in Table 4.
A study comparing 50 head and neck cancer patients found that Claude provided a more accurate diagnostic approach than ChatGPT when evaluated against multidisciplinary tumor board decisions. The study, however, did not report exact staging accuracy rates [13].
Our study demonstrated that staging results incorporating surgical pathology and clinical examination notes showed similar concordance rates with radiologic/clinical staging. However, given the 85% concordance observed for ChatGPT and lower rates for other LLMs, a cautious approach is still warranted regarding the use of such models as medical assistant tools in clinical practice. While our findings suggest that LLMs may contribute to the staging process across various cancer types—particularly in head and neck cancers—prospective validation, the development of domain-specific models, clinician training, and ethical oversight will be essential to enhance their clinical applicability.

Limitations

The limitations of this study include its retrospective design, inclusion of only non-metastatic patients, predominance of stage IV cases, and heterogeneity among the groups. The use of Turkish as the input language may have influenced model performance due to its morphologically rich structure and the limited availability of domain-specific training data [33].
While the overall distribution of disease stages reflects our real-world patient population, the relatively small number of stage I cases may have had a limited effect on the stage-specific accuracy rates. Nonetheless, this was not considered a major source of bias given the overall consistency across higher-stage groups.
A further limitation of our study is that T, N, and M components were not queried separately; instead, all results were obtained through a single-stage query. This approach may have limited our ability to assess whether stepwise prompting could improve accuracy or reduce hallucinations.
Another limitation of this study is the absence of a systematic framework for the discrete evaluation of hallucinations, out-of-scope responses, inconsistencies, and formatting errors. While such errors were implicitly captured in the overall misclassification rates and concordance analyses, they were not categorized as distinct error types. Future research should specifically address this gap by developing structured evaluation frameworks to quantify these outputs, thereby enabling a more comprehensive assessment of LLM performance in clinical oncology settings.
Finally, our study relied on a manual summarization process during case data preparation. All patient information was fully anonymized, and the summaries focused exclusively on clinical, radiological, and pathological findings relevant to TNM staging. However, no predefined structured template was used during data collection. This manual approach may carry a potential risk of human error and subjective bias. Nevertheless, this method was intentionally chosen, as in real-world clinical practice physicians typically provide LLMs with free-text case summaries rather than standardized templates.

5. Conclusions

This study addresses the critical need for the model-specific evaluation of large language models (LLMs) in oncology. While LLMs show promise for cancer staging, their performance varies substantially, requiring a cautious approach—particularly in heterogeneous diseases such as head and neck cancer, where staging depends on the integration of complex clinical, radiological, and pathological data. To this end, our work presents the first head-to-head comparison of three distinct LLMs—ChatGPT, DeepSeek, and Grok—using unstructured free-text records from real patients, thereby grounding the research in clinical reality. This methodology enabled a robust quantification of model performance across TNM components and subsites. By providing the first evidence of the clinical applicability of DeepSeek and Grok in oncology, our findings offer a foundational assessment of their strengths and limitations, underscoring the need for further model-specific studies to define the optimal role of each LLM in clinical practice.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics15182375/s1, Supplementary Table S1. Clinician-Performed Staging (distribution of T, N, M, and overall stage according to cancer localization). Supplementary Table S2. TNM Stage Accordance Between LLM and Clinician (Kappa values and pairwise comparisons). Supplementary Table S3. Concordance Rates of T, N, and Stage According to Clinical and Pathological Staging. Supplementary Table S4. Concordance Rates by Staging Methods (imaging, pathology, examination, combined). Supplementary Table S5. F1 Scores and 95% Confidence Intervals by Model and Stage. Supplementary Table S6. Concordance Rates by Gender. Supplementary Table S7. Concordance Rates by Surgical Status. Supplementary Table S8. Concordance Rates by TNM Subgroups. Supplementary Figure S1. Example of an interaction with a Large Language Model (LLM), illustrating the application of a prompt for TNM staging.

Author Contributions

Conceptualization: M.K. Data curation: H.B. Formal analysis: M.K. Investigation: M.K., H.B. and H.A.Y. Methodology: M.K. and H.A.Y. Writing—original draft: M.K. and H.A.Y. Writing—review and editing: all authors. Supervision: H.A.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical approval was obtained from the Ankara University Faculty of Medicine Human Research Ethics Committee on 13 March 2025 with the approval number 2025/168. All researchers adhered to the principles of the Declaration of Helsinki and provided their signed consent.

Informed Consent Statement

Given the retrospective design and use of fully anonymized data, the requirement for individual informed consent was waived by the Ethics Committee.

Data Availability Statement

The data analyzed in this study are not publicly available due to privacy and ethical restrictions involving patient information. Access to the data may be granted by the corresponding author upon reasonable request and with appropriate institutional approvals.

Acknowledgments

No artificial intelligence (AI) tools were used in the design, analysis, or writing of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LLMLarge language model
HNCHead and neck cancer

References

  1. Bray, F.; Laversanne, M.; Sung, H.; Ferlay, J.; Siegel, R.L.; Soerjomataram, I.; Jemal, A. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2024, 74, 229–263. [Google Scholar] [CrossRef]
  2. Machiels, J.P.; Leemans, C.R.; Golusinski, W.; Grau, C.; Licitra, L.; Gregoire, V. Squamous cell carcinoma of the oral cavity, larynx, oropharynx and hypopharynx: EHNS-ESMO-ESTRO Clinical Practice Guidelines for diagnosis, treatment and follow-up. Ann. Oncol. 2020, 31, 1462–1475. [Google Scholar] [CrossRef] [PubMed]
  3. Poon, C.S.; Stenson, K.M. Overview of the Diagnosis and Staging of Head and Neck Cancer. UptoDate (online), Fev, 2012. Available online: https://www.uptodate.com/contents/overview-of-the-diagnosis-and-staging-of-head-and-neck-cancer (accessed on 12 September 2025).
  4. Amin, M.B.; Edge, S.B.; Greene, F.L.; Byrd, D.R.; Brookland, R.K.; Washington, M.K.; Gershenwald, J.E.; Compton, C.C.; Hess, K.R.; Sullivan, D.C.; et al. AJCC Cancer Staging Manual; Springer: Berlin/Heidelberg, Germany, 2017; Volume 1024. [Google Scholar]
  5. Naveed, H.; Khan, A.U.; Qiu, S.; Saqib, M.; Anwar, S.; Usman, M.; Akhtar, N.; Barnes, N.; Mian, A. A Comprehensive Overview of Large Language Models. ACM Trans. Intell. Syst. Technol. 2023, 16, 1–72. [Google Scholar] [CrossRef]
  6. Le, N.Q.K. Leveraging transformers-based language models in proteome bioinformatics. Proteomics 2023, 23, 2300011. [Google Scholar] [CrossRef] [PubMed]
  7. Tran, T.-O.; Le, N.Q.K. Sa-TTCA: An SVM-based approach for tumor T-cell antigen classification using features extracted from biological sequencing and natural language processing. Comput. Biol. Med. 2024, 174, 108408. [Google Scholar] [CrossRef]
  8. Gibney, E. Scientists flock to DeepSeek: How they’re using the blockbuster AI model. Nature 2025. [Google Scholar] [CrossRef]
  9. Clusmann, J.; Kolbinger, F.R.; Muti, H.S.; Carrero, Z.I.; Eckardt, J.-N.; Laleh, N.G.; Löffler, C.M.L.; Schwarzkopf, S.-C.; Unger, M.; Veldhuizen, G.P.; et al. The future landscape of large language models in medicine. Commun. Med. 2023, 3, 141. [Google Scholar] [CrossRef]
  10. Carl, N.; Schramm, F.; Haggenmüller, S.; Kather, J.N.; Hetz, M.J.; Wies, C.; Michel, M.S.; Wessels, F.; Brinker, T.J. Large language model use in clinical oncology. npj Precis. Oncol. 2024, 8, 240. [Google Scholar] [CrossRef]
  11. Chizhikova, M.; López-Úbeda, P.; Martín-Noguerol, T.; Díaz-Galiano, M.C.; Ureña-López, L.A.; Luna, A.; Martín-Valdivia, M.T. Automatic TNM staging of colorectal cancer radiology reports using pre-trained language models. Comput. Methods Programs Biomed. 2025, 259, 108515. [Google Scholar] [CrossRef] [PubMed]
  12. Baran, E.; Lee, M.; Aviv, S.; Weiss, J.; Pettengell, C.; Karam, I.; Bayley, A.; Poon, I.; Chan, K.K.W.; Parmar, A.; et al. Oropharyngeal Cancer Staging Health Record Extraction Using Artificial Intelligence. JAMA Otolaryngol. Head Neck Surg. 2024, 150, 1051–1057. [Google Scholar] [CrossRef]
  13. Schmidl, B.; Hütten, T.; Pigorsch, S.; Stögbauer, F.; Hoch, C.C.; Hussain, T.; Wollenberg, B.; Wirth, M. Assessing the use of the novel tool Claude 3 in comparison to ChatGPT 4.0 as an artificial intelligence tool in the diagnosis and therapy of primary head and neck cancer cases. Eur. Arch. Oto-Rhino-Laryngol. 2024, 281, 6099–6109. [Google Scholar] [CrossRef]
  14. Tozuka, R.; Johno, H.; Amakawa, A.; Sato, J.; Muto, M.; Seki, S.; Komaba, A.; Onishi, H. Application of NotebookLM, a large language model with retrieval-augmented generation, for lung cancer staging. Jpn. J. Radiol. 2025, 43, 706–712. [Google Scholar] [CrossRef] [PubMed]
  15. OpenAi, ChatGPT (GPT-4O). 2024. Available online: www.chatgpt.com (accessed on 12 September 2025).
  16. DeepSeek. DeepSeek-V3: An Advanced AI Language Model; DeepSeek: Hangzhou, China, 2023; Available online: https://www.deepseek.com/ (accessed on 12 September 2025).
  17. xAi. Grok 3 Beta—The Age of Reasoning Agents. 2025. Available online: www.grok.com (accessed on 12 September 2025).
  18. IBM Corp. IBM SPSS Statistics; IBM Corp.: Armonk, NY, USA, 2019. [Google Scholar]
  19. Kolla, L.; Parikh, R.B. Uses and Limitations of Artificial Intelligence for Oncology. Cancer 2024, 130, 2101–2107. [Google Scholar] [CrossRef]
  20. Ahmad, M.A.; Yaramis, I.; Roy, T.D. Creating trustworthy llms: Dealing with hallucinations in healthcare ai. arXiv 2023, arXiv:2311.01463. [Google Scholar] [CrossRef]
  21. Benary, M.; Wang, X.D.; Schmidt, M.; Soll, D.; Hilfenhaus, G.; Nassir, M.; Sigler, C.; Knödler, M.; Keller, U.; Beule, D.; et al. Leveraging Large Language Models for Decision Support in Personalized Oncology. JAMA Netw. Open 2023, 6, e2343689. [Google Scholar] [CrossRef]
  22. Wei, K.; Fritz, C.; Rajasekaran, K. Answering head and neck cancer questions: An assessment of ChatGPT responses. Am. J. Otolaryngol. 2024, 45, 104085. [Google Scholar] [CrossRef] [PubMed]
  23. Marchi, F.; Bellini, E.; Iandelli, A.; Sampieri, C.; Peretti, G. Exploring the landscape of AI-assisted decision-making in head and neck cancer treatment: A comparative analysis of NCCN guidelines and ChatGPT responses. Eur. Arch. Oto-Rhino-Laryngol. 2024, 281, 2123–2136. [Google Scholar] [CrossRef] [PubMed]
  24. Lechien, J.R.; Chiesa-Estomba, C.-M.; Baudouin, R.; Hans, S. Accuracy of ChatGPT in head and neck oncological board decisions: Preliminary findings. Eur. Arch. Otorhinolaryngol. 2024, 281, 2105–2114. [Google Scholar] [CrossRef]
  25. Davis, R.J.; Ayo-Ajibola, O.; Lin, M.E.; Swanson, M.S.; Chambers, T.N.; Kwon, D.I.; Kokot, N.C. Evaluation of Oropharyngeal Cancer Information from Revolutionary Artificial Intelligence Chatbot. Laryngoscope 2024, 134, 2252–2257. [Google Scholar] [CrossRef]
  26. Choi, J.; Kim, J.W.; Lee, Y.S.; Tae, J.H.; Choi, S.Y.; Chang, I.H.; Kim, J.H. Availability of ChatGPT to provide medical information for patients with kidney cancer. Sci. Rep. 2024, 14, 1542. [Google Scholar] [CrossRef]
  27. Erdat, E.C.; Kavak, E.E. Benchmarking LLM chatbots’ oncological knowledge with the Turkish Society of Medical Oncology’s annual board examination questions. BMC Cancer 2025, 25, 197. [Google Scholar] [CrossRef] [PubMed]
  28. National Comprehensive Cancer Network. NCCN Clinical Practice Guidelines in Oncology: Head and Neck Cancers (Version 2.2025); National Comprehensive Cancer Network: Montgomery, PA, USA, 2025. [Google Scholar]
  29. Bossi, P.; Chan, A.; Licitra, L.; Trama, A.; Orlandi, E.; Hui, E.; Halámková, J.; Mattheis, S.; Baujat, B.; Hardillo, J.; et al. Nasopharyngeal carcinoma: ESMO-EURACAN Clinical Practice Guidelines for diagnosis, treatment and follow-up(†). Ann. Oncol. 2021, 32, 452–465. [Google Scholar] [CrossRef] [PubMed]
  30. Change, C.-H.; Lucas, M.M.; Lu-Yao, G.; Yang, C.C. Classifying cancer stage with open-source clinical large language models. In Proceedings of the 2024 IEEE 12th International Conference on Healthcare Informatics (ICHI), Orlando, FL, USA, 3–6 June 2024. [Google Scholar]
  31. Lee, J.E.; Park, K.-S.; Kim, Y.-H.; Song, H.-C.; Park, B.; Jeong, Y.J. Lung Cancer Staging Using Chest CT and FDG PET/CT Free-Text Reports: Comparison Among Three ChatGPT Large Language Models and Six Human Readers of Varying Experience. AJR Am. J. Roentgenol. 2024, 223, e2431696. [Google Scholar] [CrossRef]
  32. Huang, J.; Yang, D.M.; Rong, R.; Nezafati, K.; Treager, C.; Chi, Z.; Wang, S.; Cheng, X.; Guo, Y.; Klesse, L.J.; et al. A critical assessment of using ChatGPT for extracting structured data from clinical notes. npj Digit. Med. 2024, 7, 106. [Google Scholar] [CrossRef]
  33. Qin, L.; Chen, Q.; Zhou, Y.; Chen, Z.; Li, Y.; Liao, L.; Li, M.; Che, W.; Yu, P.S. A survey of multilingual large language models. Patterns 2025, 6, 101118. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Workflow of the study.
Figure 1. Workflow of the study.
Diagnostics 15 02375 g001
Figure 2. Agreement between clinicians and LLM-based models across staging systems: kappa heatmap.
Figure 2. Agreement between clinicians and LLM-based models across staging systems: kappa heatmap.
Diagnostics 15 02375 g002
Figure 3. F1 scores of models for TNM and overall stage classification.
Figure 3. F1 scores of models for TNM and overall stage classification.
Diagnostics 15 02375 g003
Table 1. Demographic and Clinical Characteristics of patients.
Table 1. Demographic and Clinical Characteristics of patients.
VariablesN (%)
Gender
Male151 (74.8)
Female51 (25.2)
Smoking
Active75 (37.1)
Exsmoker73 (36.1)
Never51 (25.2)
Age, years (median)Median (SD)
Total58.35 (14.35)
Male59.37 (13.16)
Female52.86(15.67)
Smoking Status
Never54 (26.7)
Current75 (37.1)
Ex73 (36.1)
Localization
Larynx92 (45.5)
Hypopharynx13 (6.4)
Oral Cavity/Oropharynx46 (22.8)
Nasopharynx 43 (21.3)
Salivary1 (0.5)
Nasal cavity and sinuses7 (3.5)
Definitive Surgery
Performed80 (39.6)
Not Performed122 (60.4)
Diagnostic Method
Imaging Only49 (24.3)
Pathology Only3 (1.5)
Clinical Examination Note1 (0.5)
All Methods Combined149 (73.8)
Stages
111 (5.4)
223 (11.4)
358 (28.7)
4110 (54.4)
Table 2. TNM Stage Concordance Between Large Language Model and Clinician.
Table 2. TNM Stage Concordance Between Large Language Model and Clinician.
LLM T Stage n (%)N Stage n (%)M Stage n (%)TNM Stagen (%)p
ChatGPT159 (78.7)176 (87.1)202 (100)173 (85.6)p1 < 0.001
p2 = 0.077
p3 = 0.009
p4 < 0.001
DeepSeek139 (68.8)135 (66.8)202(100)136 (67.3)
Grok130 (64.4)162 (80.2)202 (100)152 (75.2)
p1 ChatGPTstage vs. DeepSeekstage, p2 DeepSeekstage vs. Grokstage, p3 ChatGPTstage vs. Grokstage, p4 collectiveS. Abbreviations: LLM = Large Language ModelBonferroni correction (α = 0.05/3 = 0.0167).
Table 3. Cohen’s kappa Values between Clinicians and LLM Staging.
Table 3. Cohen’s kappa Values between Clinicians and LLM Staging.
LLMLarynx
n (%)
κ (sd)
Hypopharynx n (%)
κ (sd)
Oral Cavity/Oropharynx n (%)
κ (sd)
Nasopharynx n (%)
κ (sd)
Salivary
n (%)
κ (sd)
Nasal Cavity and Sinuses
n (%)
κ (sd)
Overall
n (%)
κ (sd)
ChatGPT77 (83.7)11 (84.6)41 (89.1)38 (88.4)1 (100)5 (71.4)173 (85.6)
0.778 (0.06) 0.743 (0.19)0.852 (0.07)0.791 (0.09)-0.553 (0.34)0.797 (0.04)
DeepSeek60 (65.2)8 (61.5)32 (69.6)29 (67.4)1 (100)6 (85.7)136 (67.3)
0.518 (0.09)0.268 (0.27) 0.580 (0.12)0.580 (0.12)-0.611 (0.32)0.522 (0.06)
Grok67 (72.8)12 (93.1)35 (76.1)32 (74.4)1 (100)5 (71.4)152 (75.2)
0.596 (0.08)0.687 (0.21)0.671 (0.11)0.497 (0.13)--0.614 (0.05)
Abbreviations: LLM = Large language model, SD = standard deviation, k = kappa value.
Table 4. Summary of the literature on large language models (LLMs) in oncology and cancer staging.
Table 4. Summary of the literature on large language models (LLMs) in oncology and cancer staging.
Study (Year)Cancer Type/SettingLLM EvaluatedTaskKey Findings
Tozuka et al., 2025 [14]Lung cancer (fictional cases, n = 100)NotebookLM (RAG-enhanced)TNM stagingAccuracy of ~76–86%, demonstrated feasibility of staging from reports
Chizhikova et al., 2025 [11]Colorectal cancerPre-trained LLMs (BERT, RoBERTa)Radiology report TNM extractionMacro F1 scores of 0.7464, 0.8792, and 0.6776 for T, N, and M staging, respectively
Baran et al., 2024 [12]Oropharyngeal cancerAI-based NLP (health record extraction)Automated stagingT: 55.9%, N: 56.0%, M: 87.6%, and p16: 92.1%; local vs. advanced: 80.7%
Schmidl et al., 2024 [13]Head and neck cancer (n = 50)Claude vs. ChatGPTDiagnosis, therapy recommendationsOPSCC surgery: ChatGPT: 72.2% vs Claude: 27.8%
Lee et al., 2024 [31]Lung cancerChatGPT (3 versions)Staging using CT/PET reportsGPT-4o: 74.1%; GPT-4: 70.1%; and GPT-3.5: 57.4%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kayaalp, M.; Bölek, H.; Yaşar, H.A. Confirmation of Large Language Models in Head and Neck Cancer Staging. Diagnostics 2025, 15, 2375. https://doi.org/10.3390/diagnostics15182375

AMA Style

Kayaalp M, Bölek H, Yaşar HA. Confirmation of Large Language Models in Head and Neck Cancer Staging. Diagnostics. 2025; 15(18):2375. https://doi.org/10.3390/diagnostics15182375

Chicago/Turabian Style

Kayaalp, Mehmet, Hatice Bölek, and Hatime Arzu Yaşar. 2025. "Confirmation of Large Language Models in Head and Neck Cancer Staging" Diagnostics 15, no. 18: 2375. https://doi.org/10.3390/diagnostics15182375

APA Style

Kayaalp, M., Bölek, H., & Yaşar, H. A. (2025). Confirmation of Large Language Models in Head and Neck Cancer Staging. Diagnostics, 15(18), 2375. https://doi.org/10.3390/diagnostics15182375

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop