Skip to Content
  • Tracked for
    Impact Factor
  • 3.4
    CiteScore
  • 20 days
    Time to First Decision

BioMedInformatics

BioMedInformatics is an international, peer-reviewed, open access journal on all areas of biomedical informatics, as well as computational biology and medicine, published bimonthly online by MDPI.

Get Alerted

Add your email address to receive forthcoming issues of this journal.

All Articles (346)

Background: Parkinson’s disease is a progressive neurodegenerative disorder. Early PD detection plays a key role in effective therapy. Electroencephalography is a neuroimaging technique used to analyze brain abnormalities, such as those seen in patients with PD. However, the complex nature of EEG data requires advanced signal processing and classification methods. Methods: This study systematically evaluates three time-frequency (TF) representation techniques, namely discrete wavelet transform (DWT), continuous wavelet transform (CWT), and synchrosqueezing transform (SST), along with four pretrained convolutional neural network architectures for EEG-based PD detection. The experiments were performed using the San Diego dataset. Image-wise and subject-wise 5-fold cross-validation were employed to assess performance and generalization capability. Results: CWT and SST consistently outperform DWT across all evaluated architectures in image-wise CV evaluation. At the image-wise level, the CWT-EfficientNet-B0 model achieved 97.28% accuracy for HC vs. PD-OFF classification, while SST-EfficientNet-B0 reached 97.26% accuracy for HC vs. PD-ON classification. In subject-wise evaluation, acceptable accuracies of up to 84% were achieved, indicating the ability of the framework in learning PD patterns for unseen subjects. Conclusions: These findings demonstrate that the choice of TF representation has a strong impact on classification performance and that lightweight CNN architectures can achieve high image-wise accuracy with reduced computational cost.

9 March 2026

Workflow of the proposed EEG-based model for detecting Parkinson’s disease.

Artificial Intelligence in Corneal Drug Delivery Systems

  • Amirhosein Panjipour,
  • Soheil Sojdeh and
  • Ali R. Djalilian
  • + 1 author

Conventional topical therapy for corneal and anterior segment diseases is limited by rapid tear clearance and multilayer corneal barriers, resulting in low bioavailability and the need for frequent dosing. Artificial intelligence (AI) is emerging as a complementary approach that learns quantitative relationships between molecular structure, formulation variables, and ocular performance. In corneal drug delivery, machine learning models have been used to optimize multicomponent formulations and processing conditions; predict key quality attributes such as particle size, zeta potential, encapsulation efficiency and release kinetics; and estimate corneal permeability, retention and ocular irritation risk, thereby reducing experimental burden and guiding safer design. AI can also be coupled with mechanistic ocular pharmacokinetic/pharmacodynamic models to translate formulation attributes into predicted tissue exposure. Finally, inverse design approaches enable the discovery of new carriers and devices, illustrated by machine learning-guided peptide carriers and smart contact lens platforms that combine sensing with on-demand drug release. Despite these advances, current datasets remain small and heterogeneous, external validation and benchmarking against conventional workflows are limited, and uncertainty quantification and interpretability must be addressed to enable clinical translation. This review summarizes corneal barriers and delivery platforms, critically evaluates where AI provides measurable value across design, characterization and performance and highlights data and validation priorities needed for trustworthy AI-enabled corneal therapeutics.

27 February 2026

BioMedInformatics is an international, peer-reviewed, open access journal that covers all areas of biomedical informatics, computational biology, and medicine. Established in 2021, the journal is now five years old and reflects the evolution of the field through its consistent thematic focus on Artificial Intelligence (AI)-driven diagnosis and prediction, with a particular emphasis on translational clinical decision support and biomedical signal and imaging analysis. Despite the predominance of AI-related topics, classical bioinformatics remains a major focus, with a particular emphasis on the discovery of biomarkers and the development of data resources. This editorial summarises this evolution, which accurately reflects the field as a whole.

25 February 2026

Background/Objectives: Software is now core infrastructure in biomedical science, yet fragmented workflows across subfields hinder reproducibility and delay the translation of data into actionable decisions. There is a critical need for a cross-disciplinary synthesis to bridge these silos and establish a unified framework for software maturity. This narrative review addresses this gap by synthesizing representative software ecosystems across three major pillars: bioinformatics, molecular modeling/simulations, and epidemiology/public health. Methods: A narrative review of articles indexed in PubMed/NCBI, Web of Science, and Scopus between 2000 and 2025 was conducted. Domain-specific terms related to bioinformatics, molecular modeling, docking, molecular dynamics, epidemiology, public health, and workflow management were combined with software- and algorithm-focused keywords. Studies describing, validating, or applying documented tools with biomedical relevance were included. Results: Across domains, mature data standards and reference resources (e.g., FASTQ, BAM/CRAM, VCF, mzML), widely adopted platforms (e.g., BLAST+ (v2.16.0, NCBI, Bethesda, MD, USA), Bioconductor (v3.20, Bioconductor Foundation, Seattle, WA, USA), AutoDock Vina (v1.2.5, Scripps Research, La Jolla, CA, USA), GROMACS (v2024.3, GROMACS Team, Stockholm, Sweden), Epi Info (v7.2.6, CDC, Atlanta, GA, USA), QGIS (v3.40, QGIS.org, Gossau, Switzerland), and increasing use of workflow engines were identified. Software pipelines routinely transform molecular and surveillance data into interpretable features supporting hypothesis generation. Conclusions: Integrated, standards-based, and validated software pipelines can shorten the path from measurement to decision in biomedicine and public health. Future progress depends on reproducibility practices, benchmarking, user-centered design, portable implementations, and responsible deployment of machine learning.

4 February 2026

News & Conferences

Issues

Open for Submission

Editor's Choice

XFacebookLinkedIn
BioMedInformatics - ISSN 2673-7426