Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (630)

Search Parameters:
Keywords = medical imaging modalities

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
45 pages, 4319 KiB  
Review
Advancements in Radiomics-Based AI for Pancreatic Ductal Adenocarcinoma
by Georgios Lekkas, Eleni Vrochidou and George A. Papakostas
Bioengineering 2025, 12(8), 849; https://doi.org/10.3390/bioengineering12080849 (registering DOI) - 6 Aug 2025
Abstract
The advancement of artificial intelligence (AI), deep learning, and radiomics has introduced novel methodologies for the detection, classification, prognosis, and treatment evaluation of pancreatic ductal adenocarcinoma (PDAC). As the integration of AI into medical imaging continues to evolve, its potential to enhance early [...] Read more.
The advancement of artificial intelligence (AI), deep learning, and radiomics has introduced novel methodologies for the detection, classification, prognosis, and treatment evaluation of pancreatic ductal adenocarcinoma (PDAC). As the integration of AI into medical imaging continues to evolve, its potential to enhance early detection, refine diagnostic precision, and optimize treatment strategies becomes increasingly evident. However, despite significant progress, various challenges remain, particularly in terms of clinical applicability, generalizability, interpretability, and integration into routine practice. Understanding the current state of research is crucial for identifying gaps in the literature and exploring opportunities for future advancements. This literature review aims to provide a comprehensive overview of the existing studies on AI applications in PDAC, with a focus on disease detection, classification, survival prediction, treatment response assessment, and radiogenomics. By analyzing the methodologies, findings, and limitations of these studies, we aim to highlight the strengths of AI-driven approaches while addressing critical gaps that hinder their clinical translation. Furthermore, this review aims to discuss future directions in the field, emphasizing the need for multi-institutional collaborations, explainable AI models, and the integration of multi-modal data to advance the role of AI in personalized medicine for PDAC. Full article
Show Figures

Figure 1

24 pages, 3788 KiB  
Review
Advances in Photoacoustic Imaging of Breast Cancer
by Yang Wu, Keer Huang, Guoxiong Chen and Li Lin
Sensors 2025, 25(15), 4812; https://doi.org/10.3390/s25154812 - 5 Aug 2025
Abstract
Breast cancer is the leading cause of cancer-related mortality among women world-wide, and early screening is critical for improving patient survival. Medical imaging plays a central role in breast cancer screening, diagnosis, and treatment monitoring. However, conventional imaging modalities—including mammography, ultrasound, and magnetic [...] Read more.
Breast cancer is the leading cause of cancer-related mortality among women world-wide, and early screening is critical for improving patient survival. Medical imaging plays a central role in breast cancer screening, diagnosis, and treatment monitoring. However, conventional imaging modalities—including mammography, ultrasound, and magnetic resonance imaging—face limitations such as low diagnostic specificity, relatively slow imaging speed, ionizing radiation exposure, and dependence on exogenous contrast agents. Photoacoustic imaging (PAI), a novel hybrid imaging technique that combines optical contrast with ultrasonic spatial resolution, has shown great promise in addressing these challenges. By revealing anatomical, functional, and molecular features of the breast tumor microenvironment, PAI offers high spatial resolution, rapid imaging, and minimal operator dependence. This review outlines the fundamental principles of PAI and systematically examines recent advances in its application to breast cancer screening, diagnosis, and therapeutic evaluation. Furthermore, we discuss the translational potential of PAI as an emerging breast imaging modality, complementing existing clinical techniques. Full article
(This article belongs to the Special Issue Optical Imaging for Medical Applications)
Show Figures

Figure 1

18 pages, 2315 KiB  
Systematic Review
Efficacy and Safety of Intravenous Thrombolysis in the Extended Time Window for Acute Ischemic Stroke: A Systematic Review and Meta-Analysis
by Lina Palaiodimou, Nikolaos M. Papageorgiou, Apostolos Safouris, Aikaterini Theodorou, Eleni Bakola, Maria Chondrogianni, Georgia Papagiannopoulou, Odysseas Kargiotis, Klearchos Psychogios, Eftihia Polyzogopoulou, Georgios Magoufis, Georgios Velonakis, Jobst Rudolf, Panayiotis Mitsias and Georgios Tsivgoulis
J. Clin. Med. 2025, 14(15), 5474; https://doi.org/10.3390/jcm14155474 - 4 Aug 2025
Viewed by 1
Abstract
Background/Objectives: While intravenous thrombolysis (IVT) is the standard treatment for acute ischemic stroke (AIS) within 4.5 h of symptom onset, many patients present beyond this time window. Recent trials suggest that IVT may be both effective and safe in selected patients treated after [...] Read more.
Background/Objectives: While intravenous thrombolysis (IVT) is the standard treatment for acute ischemic stroke (AIS) within 4.5 h of symptom onset, many patients present beyond this time window. Recent trials suggest that IVT may be both effective and safe in selected patients treated after the standard time window. Methods: We searched MEDLINE, Scopus, and ClinicalTrials.gov for randomized-controlled clinical trials (RCTs) and individual patient-data meta-analyses (IPDMs) of RCTs comparing IVT plus best medical treatment (BMT) to BMT alone in AIS patients who were last-known-well more than 4.5 h earlier. The primary efficacy outcome was a 90-day excellent functional outcome [modified Rankin Scale (mRS)-scores of 0–1]. Secondary efficacy outcomes included good functional outcome (mRS-scores 0–2) and reduced disability (≥1-point reduction across all mRS-strata). The primary safety outcome was symptomatic intracranial hemorrhage (sICH); secondary safety outcomes were any ICH and 3-month all-cause mortality. Subgroup analyses were performed stratified by different thrombolytics, time-windows, imaging modalities, and affected circulation. Results: Nine studies were included, comprising 1660 patients in the IVT-group and 1626 patients in the control-group. IVT significantly improved excellent functional outcome (RR = 1.24; 95%CI:1.14–1.34; I2 = 0%) and good functional outcome (RR = 1.18; 95%CI:1.05–1.33; I2 = 70%). IVT was associated with increased odds of reduced disability (common OR = 1.3; 95%CI:1.15–1.46; I2 = 0%) and increased risk of sICH (RR = 2.75; 95%CI:1.49–5.05; I2 = 0%). The rates of any ICH and all-cause mortality were similar between the two groups. No significant subgroup differences were documented. Conclusions: IVT in the extended time window improved functional outcomes without increasing mortality, despite a higher rate of sICH. Full article
(This article belongs to the Special Issue Ischemic Stroke: Diagnosis and Treatment)
Show Figures

Figure 1

29 pages, 639 KiB  
Review
Functional Pancreatic Neuroendocrine Neoplasms: An Overview
by Ethan A. Mills, Beckey P. DeLucia, Colton D. Wayne, Taylor H. Jacobs, Gail E. Besner and Siddharth Narayanan
Endocrines 2025, 6(3), 38; https://doi.org/10.3390/endocrines6030038 - 1 Aug 2025
Viewed by 493
Abstract
Pancreatic neuroendocrine neoplasms (PNENs) are a diverse group of rare tumor subtypes, representing less than 2% of all pancreatic tumors. Often detected late in the clinical course, they are associated with high rates of morbidity and mortality. Hereditary syndromes such as multiple endocrine [...] Read more.
Pancreatic neuroendocrine neoplasms (PNENs) are a diverse group of rare tumor subtypes, representing less than 2% of all pancreatic tumors. Often detected late in the clinical course, they are associated with high rates of morbidity and mortality. Hereditary syndromes such as multiple endocrine neoplasia type-1 and von Hippel–Lindau are associated with the development of PNENs, although only a small portion of total tumors have a genetic basis. This review aims to explore the recent advances in laboratory diagnostics, imaging modalities, medical management, and surgical approaches to hormone-producing PNENs (including some common, less common, and some rare subtypes), with the goal of assisting physicians in the integration of evidence-based information into their practice. Full article
Show Figures

Figure 1

23 pages, 5770 KiB  
Article
Assessment of Influencing Factors and Robustness of Computable Image Texture Features in Digital Images
by Diego Andrade, Howard C. Gifford and Mini Das
Tomography 2025, 11(8), 87; https://doi.org/10.3390/tomography11080087 (registering DOI) - 31 Jul 2025
Viewed by 129
Abstract
Background/Objectives: There is significant interest in using texture features to extract hidden image-based information. In medical imaging applications using radiomics, AI, or personalized medicine, the quest is to extract patient or disease specific information while being insensitive to other system or processing variables. [...] Read more.
Background/Objectives: There is significant interest in using texture features to extract hidden image-based information. In medical imaging applications using radiomics, AI, or personalized medicine, the quest is to extract patient or disease specific information while being insensitive to other system or processing variables. While we use digital breast tomosynthesis (DBT) to show these effects, our results would be generally applicable to a wider range of other imaging modalities and applications. Methods: We examine factors in texture estimation methods, such as quantization, pixel distance offset, and region of interest (ROI) size, that influence the magnitudes of these readily computable and widely used image texture features (specifically Haralick’s gray level co-occurrence matrix (GLCM) textural features). Results: Our results indicate that quantization is the most influential of these parameters, as it controls the size of the GLCM and range of values. We propose a new multi-resolution normalization (by either fixing ROI size or pixel offset) that can significantly reduce quantization magnitude disparities. We show reduction in mean differences in feature values by orders of magnitude; for example, reducing it to 7.34% between quantizations of 8–128, while preserving trends. Conclusions: When combining images from multiple vendors in a common analysis, large variations in texture magnitudes can arise due to differences in post-processing methods like filters. We show that significant changes in GLCM magnitude variations may arise simply due to the filter type or strength. These trends can also vary based on estimation variables (like offset distance or ROI) that can further complicate analysis and robustness. We show pathways to reduce sensitivity to such variations due to estimation methods while increasing the desired sensitivity to patient-specific information such as breast density. Finally, we show that our results obtained from simulated DBT images are consistent with what we see when applied to clinical DBT images. Full article
Show Figures

Figure 1

16 pages, 1308 KiB  
Review
Multimodality Imaging in Aldosterone-Induced Cardiomyopathy: Early Detection and Prognostic Implications
by Francesca Zoccatelli, Gabriele Costa, Matteo Merlo, Francesca Pizzolo, Simonetta Friso and Luigi Marzano
Diagnostics 2025, 15(15), 1896; https://doi.org/10.3390/diagnostics15151896 - 29 Jul 2025
Viewed by 407
Abstract
Primary aldosteronism (PA), the most common cause of secondary hypertension, is increasingly recognized as an independent driver of adverse cardiac remodeling, mediated through mechanisms beyond elevated blood pressure alone. Chronic aldosterone excess leads to myocardial fibrosis, left ventricular hypertrophy, and diastolic dysfunction via [...] Read more.
Primary aldosteronism (PA), the most common cause of secondary hypertension, is increasingly recognized as an independent driver of adverse cardiac remodeling, mediated through mechanisms beyond elevated blood pressure alone. Chronic aldosterone excess leads to myocardial fibrosis, left ventricular hypertrophy, and diastolic dysfunction via mineralocorticoid receptor activation, oxidative stress, inflammation, and extracellular matrix dysregulation. These changes culminate in a distinct cardiomyopathy phenotype, often underrecognized in early stages. Multimodality cardiac imaging, led primarily by conventional and speckle-tracking echocardiography, and complemented by exploratory cardiac magnetic resonance (CMR) techniques such as T1 mapping and late gadolinium enhancement, enables non-invasive assessment of structural, functional, and tissue-level changes in aldosterone-mediated myocardial damage. While numerous studies have established the diagnostic and prognostic relevance of imaging in PA, several gaps remain. Specifically, the relative sensitivity of different modalities in detecting subclinical myocardial changes, the long-term prognostic significance of imaging biomarkers, and the differential impact of adrenalectomy versus medical therapy on cardiac reverse remodeling require further clarification. Moreover, the lack of standardized imaging-based criteria for defining and monitoring PA-related cardiomyopathy hinders widespread clinical implementation. This narrative review aims to synthesize current knowledge on the pathophysiological mechanisms of aldosterone-induced cardiac remodeling, delineate the strengths and limitations of existing imaging modalities, and critically evaluate the comparative effects of surgical and pharmacologic interventions. Emphasis is placed on early detection strategies, identification of imaging biomarkers with prognostic utility, and integration of multimodal imaging into clinical decision-making pathways. By outlining current evidence and highlighting key unmet needs, this review provides a framework for future research aimed at advancing personalized care and improving cardiovascular outcomes in patients with PA. Full article
(This article belongs to the Section Pathology and Molecular Diagnostics)
Show Figures

Graphical abstract

23 pages, 3506 KiB  
Article
Evaluation of Vision Transformers for Multi-Organ Tumor Classification Using MRI and CT Imaging
by Óscar A. Martín and Javier Sánchez
Electronics 2025, 14(15), 2976; https://doi.org/10.3390/electronics14152976 - 25 Jul 2025
Viewed by 225
Abstract
Using neural networks has become the standard technique for medical diagnostics, especially in cancer detection and classification. This work evaluates the performance of Vision Transformer architectures, including Swin Transformer and MaxViT, for several datasets of magnetic resonance imaging (MRI) and computed tomography (CT) [...] Read more.
Using neural networks has become the standard technique for medical diagnostics, especially in cancer detection and classification. This work evaluates the performance of Vision Transformer architectures, including Swin Transformer and MaxViT, for several datasets of magnetic resonance imaging (MRI) and computed tomography (CT) scans. We used three training sets of images with brain, lung, and kidney tumors. Each dataset included different classification labels, from brain gliomas and meningiomas to benign and malignant lung conditions and kidney anomalies such as cysts and cancers. This work aims to analyze the behavior of the neural networks in each dataset and the benefits of combining different image modalities and tumor classes. We designed several experiments by fine-tuning the models on combined and individual datasets. The results revealed that the Swin Transformer achieved the highest accuracy, with an average of 99.0% on single datasets and reaching 99.43% on the combined dataset. This research highlights the adaptability of Transformer-based models to various human organs and image modalities. The main contribution lies in evaluating multiple ViT architectures across multi-organ tumor datasets, demonstrating their generalization to multi-organ classification. Integrating these models across diverse datasets could mark a significant advance in precision medicine, paving the way for more efficient healthcare solutions. Full article
(This article belongs to the Special Issue Convolutional Neural Networks and Vision Applications, 4th Edition)
Show Figures

Figure 1

28 pages, 4702 KiB  
Article
Clinical Failure of General-Purpose AI in Photographic Scoliosis Assessment: A Diagnostic Accuracy Study
by Cemre Aydin, Ozden Bedre Duygu, Asli Beril Karakas, Eda Er, Gokhan Gokmen, Anil Murat Ozturk and Figen Govsa
Medicina 2025, 61(8), 1342; https://doi.org/10.3390/medicina61081342 - 25 Jul 2025
Viewed by 348
Abstract
Background and Objectives: General-purpose multimodal large language models (LLMs) are increasingly used for medical image interpretation despite lacking clinical validation. This study evaluates the diagnostic reliability of ChatGPT-4o and Claude 2 in photographic assessment of adolescent idiopathic scoliosis (AIS) against radiological standards. This [...] Read more.
Background and Objectives: General-purpose multimodal large language models (LLMs) are increasingly used for medical image interpretation despite lacking clinical validation. This study evaluates the diagnostic reliability of ChatGPT-4o and Claude 2 in photographic assessment of adolescent idiopathic scoliosis (AIS) against radiological standards. This study examines two critical questions: whether families can derive reliable preliminary assessments from LLMs through analysis of clinical photographs and whether LLMs exhibit cognitive fidelity in their visuospatial reasoning capabilities for AIS assessment. Materials and Methods: A prospective diagnostic accuracy study (STARD-compliant) analyzed 97 adolescents (74 with AIS and 23 with postural asymmetry). Standardized clinical photographs (nine views/patient) were assessed by two LLMs and two orthopedic residents against reference radiological measurements. Primary outcomes included diagnostic accuracy (sensitivity/specificity), Cobb angle concordance (Lin’s CCC), inter-rater reliability (Cohen’s κ), and measurement agreement (Bland–Altman LoA). Results: The LLMs exhibited hazardous diagnostic inaccuracy: ChatGPT misclassified all non-AIS cases (specificity 0% [95% CI: 0.0–14.8]), while Claude 2 generated 78.3% false positives. Systematic measurement errors exceeded clinical tolerance: ChatGPT overestimated thoracic curves by +10.74° (LoA: −21.45° to +42.92°), exceeding tolerance by >800%. Both LLMs showed inverse biomechanical concordance in thoracolumbar curves (CCC ≤ −0.106). Inter-rater reliability fell below random chance (ChatGPT κ = −0.039). Universal proportional bias (slopes ≈ −1.0) caused severe curve underestimation (e.g., 10–15° error for 50° deformities). Human evaluators demonstrated superior bias control (0.3–2.8° vs. 2.6–10.7°) but suboptimal specificity (21.7–26.1%) and hazardous lumbar concordance (CCC: −0.123). Conclusions: General-purpose LLMs demonstrate clinically unacceptable inaccuracy in photographic AIS assessment, contraindicating clinical deployment. Catastrophic false positives, systematic measurement errors exceeding tolerance by 480–1074%, and inverse diagnostic concordance necessitate urgent regulatory safeguards under frameworks like the EU AI Act. Neither LLMs nor photographic human assessment achieve reliability thresholds for standalone screening, mandating domain-specific algorithm development and integration of 3D modalities. Full article
(This article belongs to the Special Issue Diagnosis and Treatment of Adolescent Idiopathic Scoliosis)
Show Figures

Figure 1

14 pages, 2935 KiB  
Article
Deep Learning-Based Differentiation of Vertebral Body Lesions on Magnetic Resonance Imaging
by Hüseyin Er, Murat Tören, Berkutay Asan, Esat Kaba and Mehmet Beyazal
Diagnostics 2025, 15(15), 1862; https://doi.org/10.3390/diagnostics15151862 - 24 Jul 2025
Viewed by 364
Abstract
Objectives: Spinal diseases are commonly encountered health problems with a wide spectrum. In addition to degenerative changes, other common spinal pathologies include metastases and compression fractures. Benign tumors like hemangiomas and infections such as spondylodiscitis are also frequently observed. Although magnetic resonance imaging [...] Read more.
Objectives: Spinal diseases are commonly encountered health problems with a wide spectrum. In addition to degenerative changes, other common spinal pathologies include metastases and compression fractures. Benign tumors like hemangiomas and infections such as spondylodiscitis are also frequently observed. Although magnetic resonance imaging (MRI) is considered the gold standard in diagnostic imaging, the morphological similarities of lesions can pose significant challenges in differential diagnoses. In recent years, the use of artificial intelligence applications in medical imaging has become increasingly widespread. In this study, we aim to detect and classify vertebral body lesions using the YOLO-v8 (You Only Look Once, version 8) deep learning architecture. Materials and Methods: This study included MRI data from 235 patients with vertebral body lesions. The dataset comprised sagittal T1- and T2-weighted sequences. The diagnostic categories consisted of acute compression fractures, metastases, hemangiomas, atypical hemangiomas, and spondylodiscitis. For automated detection and classification of vertebral lesions, the YOLOv8 deep learning model was employed. Following image standardization and data augmentation, a total of 4179 images were generated. The dataset was randomly split into training (80%) and validation (20%) subsets. Additionally, an independent test set was constructed using MRI images from 54 patients who were not included in the training or validation phases to evaluate the model’s performance. Results: In the test, the YOLOv8 model achieved classification accuracies of 0.84 and 0.85 for T1- and T2-weighted MRI sequences, respectively. Among the diagnostic categories, spondylodiscitis had the highest accuracy in the T1 dataset (0.94), while acute compression fractures were most accurately detected in the T2 dataset (0.93). Hemangiomas exhibited the lowest classification accuracy in both modalities (0.73). The F1 scores were calculated as 0.83 for T1-weighted and 0.82 for T2-weighted sequences at optimal confidence thresholds. The model’s mean average precision (mAP) 0.5 values were 0.82 for T1 and 0.86 for T2 datasets, indicating high precision in lesion detection. Conclusions: The YOLO-v8 deep learning model we used demonstrates effective performance in distinguishing vertebral body metastases from different groups of benign pathologies. Full article
Show Figures

Figure 1

15 pages, 1758 KiB  
Article
Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence
by Sahar Moradizeyveh, Ambreen Hanif, Sidong Liu, Yuankai Qi, Amin Beheshti and Antonio Di Ieva
Sensors 2025, 25(15), 4575; https://doi.org/10.3390/s25154575 - 24 Jul 2025
Viewed by 245
Abstract
Interpreting diagnostic imaging and identifying clinically relevant features remain challenging tasks, particularly for novice radiologists who often lack structured guidance and expert feedback. To bridge this gap, we propose an Eye-Gaze Guided Multimodal Fusion framework that leverages expert eye-tracking data to enhance learning [...] Read more.
Interpreting diagnostic imaging and identifying clinically relevant features remain challenging tasks, particularly for novice radiologists who often lack structured guidance and expert feedback. To bridge this gap, we propose an Eye-Gaze Guided Multimodal Fusion framework that leverages expert eye-tracking data to enhance learning and decision-making in medical image interpretation. By integrating chest X-ray (CXR) images with expert fixation maps, our approach captures radiologists’ visual attention patterns and highlights regions of interest (ROIs) critical for accurate diagnosis. The fusion model utilizes a shared backbone architecture to jointly process image and gaze modalities, thereby minimizing the impact of noise in fixation data. We validate the system’s interpretability using Gradient-weighted Class Activation Mapping (Grad-CAM) and assess both classification performance and explanation alignment with expert annotations. Comprehensive evaluations, including robustness under gaze noise and expert clinical review, demonstrate the framework’s effectiveness in improving model reliability and interpretability. This work offers a promising pathway toward intelligent, human-centered AI systems that support both diagnostic accuracy and medical training. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

20 pages, 688 KiB  
Article
Multi-Modal AI for Multi-Label Retinal Disease Prediction Using OCT and Fundus Images: A Hybrid Approach
by Amina Zedadra, Mahmoud Yassine Salah-Salah, Ouarda Zedadra and Antonio Guerrieri
Sensors 2025, 25(14), 4492; https://doi.org/10.3390/s25144492 - 19 Jul 2025
Viewed by 542
Abstract
Ocular diseases can significantly affect vision and overall quality of life, with diagnosis often being time-consuming and dependent on expert interpretation. While previous computer-aided diagnostic systems have focused primarily on medical imaging, this paper proposes VisionTrack, a multi-modal AI system for predicting multiple [...] Read more.
Ocular diseases can significantly affect vision and overall quality of life, with diagnosis often being time-consuming and dependent on expert interpretation. While previous computer-aided diagnostic systems have focused primarily on medical imaging, this paper proposes VisionTrack, a multi-modal AI system for predicting multiple retinal diseases, including Diabetic Retinopathy (DR), Age-related Macular Degeneration (AMD), Diabetic Macular Edema (DME), drusen, Central Serous Retinopathy (CSR), and Macular Hole (MH), as well as normal cases. The proposed framework integrates a Convolutional Neural Network (CNN) for image-based feature extraction, a Graph Neural Network (GNN) to model complex relationships among clinical risk factors, and a Large Language Model (LLM) to process patient medical reports. By leveraging diverse data sources, VisionTrack improves prediction accuracy and offers a more comprehensive assessment of retinal health. Experimental results demonstrate the effectiveness of this hybrid system, highlighting its potential for early detection, risk assessment, and personalized ophthalmic care. Experiments were conducted using two publicly available datasets, RetinalOCT and RFMID, which provide diverse retinal imaging modalities: OCT images and fundus images, respectively. The proposed multi-modal AI system demonstrated strong performance in multi-label disease prediction. On the RetinalOCT dataset, the model achieved an accuracy of 0.980, F1-score of 0.979, recall of 0.978, and precision of 0.979. Similarly, on the RFMID dataset, it reached an accuracy of 0.989, F1-score of 0.881, recall of 0.866, and precision of 0.897. These results confirm the robustness, reliability, and generalization capability of the proposed approach across different imaging modalities. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

33 pages, 15612 KiB  
Article
A Personalized Multimodal Federated Learning Framework for Skin Cancer Diagnosis
by Shuhuan Fan, Awais Ahmed, Xiaoyang Zeng, Rui Xi and Mengshu Hou
Electronics 2025, 14(14), 2880; https://doi.org/10.3390/electronics14142880 - 18 Jul 2025
Viewed by 339
Abstract
Skin cancer is one of the most prevalent forms of cancer worldwide, and early and accurate diagnosis critically impacts patient outcomes. Given the sensitive nature of medical data and its fragmented distribution across institutions (data silos), privacy-preserving collaborative learning is essential to enable [...] Read more.
Skin cancer is one of the most prevalent forms of cancer worldwide, and early and accurate diagnosis critically impacts patient outcomes. Given the sensitive nature of medical data and its fragmented distribution across institutions (data silos), privacy-preserving collaborative learning is essential to enable knowledge-sharing without compromising patient confidentiality. While federated learning (FL) offers a promising solution, existing methods struggle with heterogeneous and missing modalities across institutions, which reduce the diagnostic accuracy. To address these challenges, we propose an effective and flexible Personalized Multimodal Federated Learning framework (PMM-FL), which enables efficient cross-client knowledge transfer while maintaining personalized performance under heterogeneous and incomplete modality conditions. Our study contains three key contributions: (1) A hierarchical aggregation strategy that decouples multi-module aggregation from local deployment via global modular-separated aggregation and local client fine-tuning. Unlike conventional FL (which synchronizes all parameters in each round), our method adopts a frequency-adaptive synchronization mechanism, updating parameters based on their stability and functional roles. (2) A multimodal fusion approach based on multitask learning, integrating learnable modality imputation and attention-based feature fusion to handle missing modalities. (3) A custom dataset combining multi-year International Skin Imaging Collaboration(ISIC) challenge data (2018–2024) to ensure comprehensive coverage of diverse skin cancer types. We evaluate PMM-FL through diverse experiment settings, demonstrating its effectiveness in heterogeneous and incomplete modality federated learning settings, achieving 92.32% diagnostic accuracy with only a 2% drop in accuracy under 30% modality missingness, with a 32.9% communication overhead decline compared with baseline FL methods. Full article
(This article belongs to the Special Issue Multimodal Learning and Transfer Learning)
Show Figures

Figure 1

14 pages, 1509 KiB  
Article
A Multi-Modal Deep Learning Approach for Predicting Eligibility for Adaptive Radiation Therapy in Nasopharyngeal Carcinoma Patients
by Zhichun Li, Zihan Li, Sai Kit Lam, Xiang Wang, Peilin Wang, Liming Song, Francis Kar-Ho Lee, Celia Wai-Yi Yip, Jing Cai and Tian Li
Cancers 2025, 17(14), 2350; https://doi.org/10.3390/cancers17142350 - 15 Jul 2025
Viewed by 345
Abstract
Background: Adaptive radiation therapy (ART) can improve prognosis for nasopharyngeal carcinoma (NPC) patients. However, the inter-individual variability in anatomical changes, along with the resulting extension of treatment duration and increased workload for the radiologists, makes the selection of eligible patients a persistent challenge [...] Read more.
Background: Adaptive radiation therapy (ART) can improve prognosis for nasopharyngeal carcinoma (NPC) patients. However, the inter-individual variability in anatomical changes, along with the resulting extension of treatment duration and increased workload for the radiologists, makes the selection of eligible patients a persistent challenge in clinical practice. The purpose of this study was to predict eligible ART candidates prior to radiation therapy (RT) for NPC patients using a classification neural network. By leveraging the fusion of medical imaging and clinical data, this method aimed to save time and resources in clinical workflows and improve treatment efficiency. Methods: We collected retrospective data from 305 NPC patients who received RT at Hong Kong Queen Elizabeth Hospital. Each patient sample included pre-treatment computed tomographic (CT) images, T1-weighted magnetic resonance imaging (MRI) data, and T2-weighted MRI images, along with clinical data. We developed and trained a novel multi-modal classification neural network that combines ResNet-50, cross-attention, multi-scale features, and clinical data for multi-modal fusion. The patients were categorized into two labels based on their re-plan status: patients who received ART during RT treatment, as determined by the radiation oncologist, and those who did not. Results: The experimental results demonstrated that the proposed multi-modal deep prediction model outperformed other commonly used deep learning networks, achieving an area under the curve (AUC) of 0.9070. These results indicated the ability of the model to accurately classify and predict ART eligibility for NPC patients. Conclusions: The proposed method showed good performance in predicting ART eligibility among NPC patients, highlighting its potential to enhance clinical decision-making, optimize treatment efficiency, and support more personalized cancer care. Full article
Show Figures

Figure 1

22 pages, 2320 KiB  
Review
Use of Radiomics in Characterizing Tumor Hypoxia
by Mohan Huang, Helen K. W. Law and Shing Yau Tam
Int. J. Mol. Sci. 2025, 26(14), 6679; https://doi.org/10.3390/ijms26146679 - 11 Jul 2025
Viewed by 491
Abstract
Tumor hypoxia involves limited oxygen supply within the tumor microenvironment and is closely associated with aggressiveness, metastasis, and resistance to common cancer treatment modalities such as chemotherapy and radiotherapy. Traditional methodologies for hypoxia assessment, such as the use of invasive probes and clinical [...] Read more.
Tumor hypoxia involves limited oxygen supply within the tumor microenvironment and is closely associated with aggressiveness, metastasis, and resistance to common cancer treatment modalities such as chemotherapy and radiotherapy. Traditional methodologies for hypoxia assessment, such as the use of invasive probes and clinical biomarkers, are generally not very suitable for routine clinical applications. Radiomics provides a non-invasive approach to hypoxia assessment by extracting quantitative features from medical images. Thus, radiomics is important in diagnosis and the formulation of a treatment strategy for tumor hypoxia. This article discusses the various imaging techniques used for the assessment of tumor hypoxia including magnetic resonance imaging (MRI), positron emission tomography (PET), and computed tomography (CT). It introduces the use of radiomics with machine learning and deep learning for extracting quantitative features, along with its possible clinical use in hypoxic tumors. This article further summarizes the key challenges hindering the clinical translation of radiomics, including the lack of imaging standardization and the limited availability of hypoxia-labeled datasets. It also highlights the potential of integrating radiomics with multi-omics to enhance hypoxia visualization and guide personalized cancer treatment. Full article
Show Figures

Figure 1

18 pages, 1667 KiB  
Article
Multi-Task Deep Learning for Simultaneous Classification and Segmentation of Cancer Pathologies in Diverse Medical Imaging Modalities
by Maryem Rhanoui, Khaoula Alaoui Belghiti and Mounia Mikram
Onco 2025, 5(3), 34; https://doi.org/10.3390/onco5030034 - 11 Jul 2025
Viewed by 407
Abstract
Background: Clinical imaging is an important part of health care providing physicians with great assistance in patients treatment. In fact, segmentation and grading of tumors can help doctors assess the severity of the cancer at an early stage and increase the chances [...] Read more.
Background: Clinical imaging is an important part of health care providing physicians with great assistance in patients treatment. In fact, segmentation and grading of tumors can help doctors assess the severity of the cancer at an early stage and increase the chances of cure. Despite that Deep Learning for cancer diagnosis has achieved clinically acceptable accuracy, there still remains challenging tasks, especially in the context of insufficient labeled data and the subsequent need for expensive computational ressources. Objective: This paper presents a lightweight classification and segmentation deep learning model to assist in the identification of cancerous tumors with high accuracy despite the scarcity of medical data. Methods: We propose a multi-task architecture for classification and segmentation of cancerous tumors in the Brain, Skin, Prostate and lungs. The model is based on the UNet architecture with different pre-trained deep learning models (VGG 16 and MobileNetv2) as a backbone. The multi-task model is validated on relatively small datasets (slightly exceed 1200 images) that are diverse in terms of modalities (IRM, X-Ray, Dermoscopic and Digital Histopathology), number of classes, shapes, and sizes of cancer pathologies using the accuracy and dice coefficient as statistical metrics. Results: Experiments show that the multi-task approach improve the learning efficiency and the prediction accuracy for the segmentation and classification tasks, compared to training the individual models separately. The multi-task architecture reached a classification accuracy of 86%, 90%, 88%, and 87% respectively for Skin Lesion, Brain Tumor, Prostate Cancer and Pneumothorax. For the segmentation tasks we were able to achieve high precisions respectively 95%, 98% for the Skin Lesion and Brain Tumor segmentation and a 99% precise segmentation for both Prostate cancer and Pneumothorax. Proving that the multi-task solution is more efficient than single-task networks. Full article
Show Figures

Figure 1

Back to TopTop