Next Article in Journal
Class III Malocclusion in Growing Patients: Facemask vs. Functional Appliance: Experimental Study
Previous Article in Journal
Can Artificial Intelligence Improve the Appropriate Use and Decrease the Misuse of REBOA?
Previous Article in Special Issue
The Evaluation of a Deep Learning Approach to Automatic Segmentation of Teeth and Shade Guides for Tooth Shade Matching Using the SAM2 Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

AI Advancements in Healthcare: Medical Imaging and Sensing Technologies

by
Mohammed A. Al-masni
1,* and
Kanghyun Ryu
2,3,*
1
Department of Artificial Intelligence and Data Science, College of Artificial Intelligence Convergence, Sejong University, Seoul 05006, Republic of Korea
2
Intelligence and Interaction Research Center, Korea Institute of Science and Technology (KIST), Seoul 02792, Republic of Korea
3
AI-Robotics, KIST School, University of Science and Technology, Seoul 01811, Republic of Korea
*
Authors to whom correspondence should be addressed.
Bioengineering 2025, 12(10), 1026; https://doi.org/10.3390/bioengineering12101026
Submission received: 15 September 2025 / Revised: 21 September 2025 / Accepted: 23 September 2025 / Published: 26 September 2025

1. Introduction

Artificial intelligence (AI), broadly defined as algorithms capable of self-learning patterns from large-scale data, has emerged as one of the most transformative technologies in modern healthcare [1]. Advances in medical imaging and sensing have created unprecedented opportunities for data-driven analysis. Imaging modalities such as magnetic resonance imaging (MRI), computed tomography (CT), and X-ray offer detailed visualization of structural changes in the body, while sensing techniques such as electroencephalography (EEG), electrocardiography (ECG), and electromyography (EMG) provide complementary measurements of physiological activity. Together, these modalities generate complex datasets ideally suited for AI-driven modeling. When integrated with advanced computational methods, they extend far beyond traditional diagnostic roles to enable earlier disease detection, personalized treatment planning, and continuous monitoring of health trajectories.
Looking forward, the role of AI in healthcare will hinge on developing systems that are not only innovative but also trustworthy. The FUTURE-AI framework [2] highlights six guiding principles—fairness, universality, traceability, usability, robustness, and explainability—as essential for ensuring safe and reliable clinical adoption.
This Special Issue of Bioengineering, titled “AI Advancements in Healthcare: Medical Imaging and Sensing Technologies”, brings together five innovative contributions [3,4,5,6,7] that collectively represent the state of the art in this rapidly evolving field. The selected works span a broad spectrum of imaging modalities—from video silhouettes and dermoscopy to MRI, chest X-ray, and intraoral photography—and apply a diverse range of AI methodologies, including novel convolutional neural network (CNN) architectures, quantum-enhanced classifiers, interpretable brain-aging biomarkers, vision–language pipelines, and foundation segmentation models. Each article situates its technical advances within a clinically meaningful application, addressing challenges such as neurological gait assessment, dermatological cancer screening, neurodegenerative disease staging, radiological reporting, and restorative dentistry. However, it should be noted that although the theme of this Special Issue encompasses both imaging and sensing technologies, the final collection of accepted articles is focused exclusively on imaging applications. No contributions addressing sensing modalities were included among the published works, and we hope that future editions of this Special Issue will also attract submissions that highlight AI-driven advances in sensing technologies.
Collectively, these studies highlight not only the innovative use of AI technologies but also their translational potential, showing the alignment of computational advances with pressing clinical needs. Table 1 provides a comparative overview of the five articles, summarizing their clinical aims, imaging modalities and datasets, methodological innovations, validation strategies, and interpretability features.

2. Overview of Published Articles

This section provides an overview of the five contributions included in this Special Issue. Each subsection briefly summarizes the motivation of the work, the technical AI advancements presented, and the clinical implications, including considerations of feasibility and interpretability.

2.1. Functional Assessment Using Gait Dynamics

The first contribution in this Special Issue focuses on gait analysis [3] as a proxy for functional assessment in neurological disorders. The study introduces a novel method that transforms binary silhouette sequences of walking subjects into sinograms by projecting pixel intensities over a range of angular directions. This process yields motion-encoded maps in which temporal dynamics of gait are compactly represented as continuous angular patterns. These sinogram representations are subsequently processed by a 1D-CNN equipped with an assisted knowledge learning strategy, enabling the model to capture both local temporal fluctuations and global gait signatures. Evaluated on the INIT GAIT dataset [8,9], the framework demonstrated consistently high accuracy across different classification schemes, including both frame-level analysis and subject-level aggregation through majority voting. A key advantage of the method is that the sinogram representation makes the underlying gait dynamics easier to interpret, as it directly encodes angular motion in a form that clinicians can relate to observable movement patterns. By combining this representation with low-cost video input, the study demonstrates a practical pathway toward the development of scalable tools that could support neurological screening and rehabilitation in real-world settings.

2.2. Dermatological Image Analysis

The second contribution addresses the growing demand for automated computer-aided diagnosis (CAD) in dermatology, focusing on multi-class classification of skin lesions [4]. To enhance generalizability, the study draws on three widely used public datasets—PAD-UFES-20-Modified [10,11], ISIC-2018 [12], and ISIC-2019 [13]—thereby capturing a broad spectrum of lesion appearances and acquisition conditions. The proposed framework integrates a CNN with an autoencoder to extract both local image features and compact latent representations. These features are then processed by a quantum support vector machine (QSVM), which is designed to exploit high-dimensional feature spaces for complex decision boundaries.
A critical aspect of the work lies in its comprehensive preprocessing strategy, including image normalization and enhancement, which ensures consistent input quality across heterogeneous datasets. By combining classical deep learning with quantum-inspired classification, the study illustrates a promising direction for dermatological CAD systems, particularly in settings where robustness across multiple datasets is essential. Although the use of QSVM introduces challenges for interpretability, this limitation opens avenues for future research on integrating explainable AI tools to increase clinical trust.

2.3. AI Biomarkers for Neurodegenerative Diseases

The third contribution presents an innovative approach to quantifying brain aging at a regional level, with direct implications for neurodegenerative disease research and early diagnosis [5]. Building on the concept of brain-age prediction models, the authors developed the Regional Brain Aging Disparity Index (RBADI), which captures localized deviations between predicted and chronological age across distinct brain regions. The method relies on a deep learning-based brain-age model trained on a large-scale cohort from the UK Biobank [14], encompassing over sixteen thousand healthy individuals. To improve interpretability, the study employs a multi-stage Shapley value approximation, enabling the attribution of brain-age predictions to anatomically meaningful regions and therefore grounding the RBADI in established neuroanatomy.
The framework was externally validated on independent datasets from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) [15] and the Parkinson’s Progression Markers Initiative (PPMI) [16], demonstrating its applicability across multiple neurological conditions. Beyond diagnostic staging, RBADI was further linked to lifestyle and demographic factors, highlighting its potential as a population-level biomarker for understanding risk profiles. By integrating large-cohort training, external validation, and explainable AI techniques, this work illustrates how region-specific biomarkers can bridge the gap between computational models and clinically relevant insights into neurodegenerative disease progression.

2.4. Vision–Language Models for Radiological Reporting

The fourth contribution advances the integration of AI into radiological workflows by tackling the complex task of automated report generation from chest X-rays [6]. Unlike conventional pipelines that focus narrowly on classification or segmentation, this study introduces the Integrated Hierarchical Radiology Assistant System (IHRAS), an end-to-end framework designed to produce clinically meaningful textual reports. The system combines multiple AI components: a CNN for multi-label disease classification, Grad-CAM for visual explanation, SAR-Net for anatomical region segmentation, and a large language model (LLM) guided by structured prompt engineering. Importantly, the language component is aligned with established ontologies such as SNOMED CT, ensuring consistency with medical standards.
The framework is built on the large-scale NIH ChestX-ray dataset [17] and emphasizes not only diagnostic performance but also the coherence, relevance, and faithfulness of the generated reports. By fusing image classification, region-level anatomical context, and language modeling, IHRAS demonstrates how vision–language integration can move AI systems closer to supporting radiologists in daily practice. The inclusion of interpretability mechanisms, both visual (Grad-CAM) and linguistic (standardized terminology), strengthens transparency and clinical trust, positioning this work as a significant step toward practical AI adoption in radiology reporting.

2.5. Foundation Models for Dental Imaging

The final contribution in this Special Issue explores the application of foundation models to restorative dentistry, with a focus on automating tooth and shade-guide segmentation from intraoral photographs [7]. Accurate shade matching is a critical step in dental restoration, yet it remains prone to variability due to lighting conditions, subjective assessment, and manual delineation. To address this challenge, the study evaluated multiple variants of the Segment Anything Model 2 (SAM2), including tiny, small, base+, and large configurations against a conventional U-Net baseline. By fine-tuning SAM2 on a curated dataset of intraoral images, the authors demonstrated the capacity of large-scale pretrained segmentation models to adapt effectively to highly specialized clinical tasks.
A noteworthy aspect of the study is its emphasis on not only boundary precision but also color fidelity, assessed through perceptually calibrated metrics such as CIELAB and ΔE00. This dual evaluation shows the practical importance of segmentation quality for ensuring accurate shade selection in restorative workflows. While the dataset was limited to a single-center collection, the findings highlight the translational promise of adapting foundation models to dentistry, a domain where annotated data are typically scarce. By leveraging pretrained general-purpose vision models, this work points toward the more efficient development of AI tools for specialized yet clinically significant applications.

3. Conclusions

This Special Issue of Bioengineering shows the breadth and depth of current advances in artificial intelligence for healthcare, encompassing functional assessment, dermatological diagnosis, neuroimaging biomarkers, radiological reporting, and dental applications. Across these diverse domains, the featured studies demonstrate how methodological innovations, ranging from novel signal representations and hybrid learning pipelines to interpretable biomarkers, vision–language integration, and the adaptation of foundation models, can be translated into clinically meaningful outcomes. These contributions highlight that progress in healthcare AI must extend beyond improvements in accuracy to also embrace interpretability, standardization, and practical usability, ensuring that technological developments are aligned with the realities of clinical practice.

Author Contributions

Conceptualization, M.A.A.-m. and K.R.; validation, M.A.A.-m. and K.R.; formal analysis, M.A.A.-m. and K.R.; investigation, M.A.A.-m. and K.R.; writing—original draft preparation, M.A.A.-m.; writing—review and editing, K.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The editors thank the authors and reviewers for their valuable contributions to this Special Issue of Bioengineering. Special appreciation is also extended to the Managing Editor, for her dedicated support throughout the editorial process. This work was supported by the National Research Foundation of Korea (NRF) funded by the Korean government (MSIT) (No. RS-2023-00243034). This work was also supported by the Korea Institute of Science and Technology (KIST) Institutional Program under Grant No. 2E3375C, 2E33842, 2E33854.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ADNIAlzheimer’s Disease Neuroimaging Initiative
CADComputer-Aided Diagnosis
CNNConvolutional Neural Network
CTComputed Tomography
CXRChest X-ray
ECGElectrocardiography
EEGElectroencephalography
EMGElectromyography
Grad-CAMGradient-Weighted Class Activation Map
IHRASIntegrated Hierarchical Radiology Assistant System
ISICInternational Skin Imaging Collaboration
LLMLarge Language Model
MRIMagnetic Resonance Imaging
PPMIParkinson’s Progression Markers Initiative
QSVMQuantum Support Vector Machine
RBADIRegional Brain Aging Deviation Index
SAMSegment Anything Model
SAR-NetStructure-Aware Relation Network
SVMSupport Vector Machine
XAIExplainable Artificial Intelligence

References

  1. Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef] [PubMed]
  2. Lekadir, K.; Frangi, A.F.; Porras, A.R.; Glocker, B.; Cintas, C.; Langlotz, C.P.; Weicken, E.; Asselbergs, F.W.; Prior, F.; Collins, G.S.; et al. FUTURE-AI: International consensus guideline for trustworthy and deployable artificial intelligence in healthcare. BMJ 2025, 388, e081554. [Google Scholar] [CrossRef] [PubMed]
  3. Al-Masni, M.A.; Marzban, E.N.; Al-Shamiri, A.K.; Al-Antari, M.A.; Alabdulhafith, M.I.; Mahmoud, N.F.; Abdel Samee, N.; Kadah, Y.M. Gait Impairment Analysis Using Silhouette Sinogram Signals and Assisted Knowledge Learning. Bioengineering 2024, 11, 477. [Google Scholar] [CrossRef] [PubMed]
  4. Khan, A.; Sajid, M.Z.; Khan, N.A.; Youssef, A.; Abbas, Q. CAD-Skin: A Hybrid Convolutional Neural Network-Autoencoder Framework for Precise Detection and Classification of Skin Lesions and Cancer. Bioengineering 2025, 12, 326. [Google Scholar] [CrossRef] [PubMed]
  5. Wu, Y.; Sun, S.; Zhang, C.; Ma, X.; Zhu, X.; Li, Y.; Lin, L.; Fu, Z. Regional Brain Aging Disparity Index: Region-Specific Brain Aging State Index for Neurodegenerative Diseases and Chronic Disease Specificity. Bioengineering 2025, 12, 607. [Google Scholar] [CrossRef] [PubMed]
  6. Rodrigues, G.A.P.; Serrano, A.L.M.; Bispo, G.D.; Filho, G.P.R.; Goncalves, V.P.; Meneguette, R.I. IHRAS: Automated Medical Report Generation from Chest X-Rays via Classification, Segmentation, and LLMs. Bioengineering 2025, 12, 795. [Google Scholar] [CrossRef] [PubMed]
  7. Han, K.H.; Lim, J.H.; Ahn, J.-S.; Lee, K.-S. The Evaluation of a Deep Learning Approach to Automatic Segmentation of Teeth and Shade Guides for Tooth Shade Matching Using the SAM2 Algorithm. Bioengineering 2025, 12, 959. [Google Scholar] [CrossRef]
  8. Ortells, J.; Herrero-Ezquerro, M.T.; Mollineda, R.A. Vision-based gait impairment analysis for aided diagnosis. Med. Biol. Eng. Comput. 2018, 56, 1553–1564. [Google Scholar] [CrossRef] [PubMed]
  9. INIT Gait Database. Available online: https://www.vision.uji.es/gaitDB/ (accessed on 13 September 2024).
  10. Pacheco, A.G.C.; Lima, G.R.; Salomao, A.S.; Krohling, B.; Biral, I.P.; de Angelo, G.G.; Alves, F.C.R., Jr.; Esgario, J.G.M.; Simora, A.C.; Castro, P.B.C.; et al. PAD-UFES-20: A skin lesion dataset composed of patient data and clinical images collected from smartphones. Data Brief 2020, 32, 106221. [Google Scholar] [CrossRef] [PubMed]
  11. Azeem, M.; Kiani, K.; Mansouri, T.; Topping, N. SkinLesNet: Classification of Skin Lesions and Detection of Melanoma Cancer Using a Novel Multi-Layer Deep Convolutional Neural Network. Cancers 2023, 16, 108. [Google Scholar] [CrossRef] [PubMed]
  12. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv 2019, arXiv:1902.03368. [Google Scholar] [CrossRef]
  13. Hernandez-Perez, C.; Combalia, M.; Podlipnik, S.; Codella, N.C.F.; Rotemberg, V.; Halpern, A.C.; Reiter, O.; Carrera, C.; Barreiro, A.; Helba, B.; et al. BCN20000: Dermoscopic Lesions in the Wild. Sci. Data 2024, 11, 641. [Google Scholar] [CrossRef] [PubMed]
  14. UK Biobank (UKB). Available online: http://www.ukbiobank.ac.uk/register-apply/ (accessed on 10 August 2024).
  15. Alzheimer’s Disease Neuroimaging Initiative (ADNI). Available online: https://adni.loni.usc.edu/data-samples/adni-data/ (accessed on 15 April 2025).
  16. Parkinson’s Progression Markers Initiative (PPMI). Available online: https://www.ppmi-info.org/access-data-specimens/download-data (accessed on 15 April 2025).
  17. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
Table 1. A summary of the five articles published in this Special Issue, titled “AI Advancements in Healthcare: Medical Imaging and Sensing Technologies”.
Table 1. A summary of the five articles published in this Special Issue, titled “AI Advancements in Healthcare: Medical Imaging and Sensing Technologies”.
Ref.Clinical AimModality/DatasetMethodologyValidation SetupInterpretability
Al-masni et al. [3]Screening and assessment of gait abnormalities2D video silhouettes/INIT GAIT [8,9]Novel silhouette sinogram representation and 1D-CNN with assisted knowledge learningTwo schemes: frame-level classification and subject-level majority votingIntrinsic interpretability through sinogram encoding of angular motion
Khan et al. [4]Automated CAD for melanoma and other skin lesion typesDermoscopy/PAD-UFES-20-Modified [10,11], ISIC-2018 [12], ISIC-2019 [13]Hybrid CNN–autoencoder with advanced preprocessing; quantum SVM as final classifierCross-dataset benchmarking; class imbalance addressed via augmentationLimited intrinsic interpretability; potential for XAI to enhance clinical trust
Wu et al. [5]Region-level aging analysis and staging of neurodegenerationT1-weighted MRI/UK Biobank [14], ADNI [15], PPMI [16]DL-based brain-age prediction with multi-stage Shapley value approximation → RBADI biomarkerTrain/val/test split on UKB; external validation on ADNI and PPMIStrong interpretability via regional Shapley attribution; anatomically grounded RBADI
Rodrigues et al. [6]AI pipeline for diagnostic reportingChest X-ray/NIH CXR [17]CNN multi-label classifier (14 diseases) + Grad-CAM + SAR-Net anatomical segmentation + LLM (DeepSeek-R1, CRISPE prompts, SNOMED CT)Internal benchmarking across diverse demographic subgroupsMulti-level interpretability: visual explanations (Grad-CAM), anatomical context (SAR-Net), and standardized terminology (SNOMED CT)
Han et al. [7]Tooth and shade-guide segmentation for shade matchingIntraoral photographs/private datasetFine-tuned Segment Anything Model 2 (tiny/small/base+/large) vs. UNet baselineHeld-out test split; comparative evaluation across SAM2 scalesRobust interpretability via boundary precision and quantitative color fidelity (CIELAB, ΔE00)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-masni, M.A.; Ryu, K. AI Advancements in Healthcare: Medical Imaging and Sensing Technologies. Bioengineering 2025, 12, 1026. https://doi.org/10.3390/bioengineering12101026

AMA Style

Al-masni MA, Ryu K. AI Advancements in Healthcare: Medical Imaging and Sensing Technologies. Bioengineering. 2025; 12(10):1026. https://doi.org/10.3390/bioengineering12101026

Chicago/Turabian Style

Al-masni, Mohammed A., and Kanghyun Ryu. 2025. "AI Advancements in Healthcare: Medical Imaging and Sensing Technologies" Bioengineering 12, no. 10: 1026. https://doi.org/10.3390/bioengineering12101026

APA Style

Al-masni, M. A., & Ryu, K. (2025). AI Advancements in Healthcare: Medical Imaging and Sensing Technologies. Bioengineering, 12(10), 1026. https://doi.org/10.3390/bioengineering12101026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop