Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (27,018)

Search Parameters:
Keywords = image feature

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 2497 KB  
Article
Oil Spill Identification with Marine Radar Using Feature Augmentation and Improved Firefly Optimization Algorithm
by Jin Xu, Boxi Yao, Haihui Dong, Zekun Guo, Bo Xu, Yuanyuan Huang, Bo Li, Sihan Qian and Bingxin Liu
Remote Sens. 2025, 17(18), 3148; https://doi.org/10.3390/rs17183148 - 10 Sep 2025
Abstract
Oil spill accidents pose a grave threat to marine ecosystems, human economy, and public health. Consequently, expeditious and efficacious oil spill detection technology is imperative for the pollution mitigation and the health preservation in the marine environment. This study proposed a marine radar [...] Read more.
Oil spill accidents pose a grave threat to marine ecosystems, human economy, and public health. Consequently, expeditious and efficacious oil spill detection technology is imperative for the pollution mitigation and the health preservation in the marine environment. This study proposed a marine radar oil spill detection method based on Local Binary Patterns (LBP), Histogram of Oriented Gradient (HOG), and an improved Firefly Optimization Algorithm (IFA). In the stage of image pre-processing, the oil film features were significantly enhanced through three steps. The LBP features were extracted from the preprocessed image. Then, the mean filtering was used to smooth out the LBP features. Subsequently, the HOG statistical features were extracted from the filtered LBP feature map. After the feature enhancement, the oil spill regions were accurately extracted by using K-Means clustering algorithm. Next, an IFA model was used to classify oil films. Compared with traditional Firefly Optimization Algorithm (FA) algorithm, the IFA method is suitable for oil film segmentation tasks in marine radar data. The proposed method can achieve accuracy segmentation and provide a new technical path for marine oil spill monitoring. Full article
21 pages, 6526 KB  
Article
Tissue Characterization by Ultrasound: Linking Envelope Statistics with Spectral Analysis for Simultaneous Attenuation Coefficient and Scatterer Clustering Quantification
by Luis Elvira, Carla de León, Carmen Durán, Alberto Ibáñez, Montserrat Parrilla and Óscar Martínez-Graullera
Appl. Sci. 2025, 15(18), 9924; https://doi.org/10.3390/app15189924 - 10 Sep 2025
Abstract
This paper proposes the use of quantitative methods for the characterization of tissues by linking, into a single approach, ideas coming from the spectral analysis methods commonly used to determine the attenuation coefficient with the envelope statistics formulation. Initially, the Homodyned K-distribution model [...] Read more.
This paper proposes the use of quantitative methods for the characterization of tissues by linking, into a single approach, ideas coming from the spectral analysis methods commonly used to determine the attenuation coefficient with the envelope statistics formulation. Initially, the Homodyned K-distribution model used to fit data obtained from ultrasound signal envelopes was reviewed, and the necessary equations to further derive the attenuation coefficient from this model were developed. To test and discuss the performance of these methods, experimental work was conducted in phantoms. To this end, a series of tissue-mimicking materials composed of poly-vinyl alcohol (PVA) loaded with different particles (aluminium, alumina, cellulose) at varying concentrations were manufactured. A single-channel scanning system was employed to analyse these samples. It was verified that quantitative images obtained from the attenuation coefficient and from the scatterer clustering μ parameter (associated with scatterer concentration) effectively discriminate materials exhibiting similar echo envelope patterns, enhancing the information obtained in comparison with the conventional analysis based on B-scans. Additionally, the implementation of quantitative bi-parametric imaging mappings based on both the μ parameter and the attenuation coefficient, as a means to rapidly visualize results and identify areas characterized by specific acoustic features, was also proposed. Full article
(This article belongs to the Special Issue Applications of Ultrasonic Technology in Biomedical Sciences)
Show Figures

Figure 1

22 pages, 5732 KB  
Article
Explainable Transformer-Based Framework for Glaucoma Detection from Fundus Images Using Multi-Backbone Segmentation and vCDR-Based Classification
by Hind Alasmari, Ghada Amoudi and Hanan Alghamdi
Diagnostics 2025, 15(18), 2301; https://doi.org/10.3390/diagnostics15182301 - 10 Sep 2025
Abstract
Glaucoma is an eye disease caused by increased intraocular pressure (IOP) that affects the optic nerve head (ONH), leading to vision problems and irreversible blindness. Background/Objectives: Glaucoma is the second leading cause of blindness worldwide, and the number of people affected is [...] Read more.
Glaucoma is an eye disease caused by increased intraocular pressure (IOP) that affects the optic nerve head (ONH), leading to vision problems and irreversible blindness. Background/Objectives: Glaucoma is the second leading cause of blindness worldwide, and the number of people affected is increasing each year, with the number expected to reach 111.8 million by 2040. This escalating trend is alarming due to the lack of ophthalmology specialists relative to the population. This study proposes an explainable end-to-end pipeline for automated glaucoma diagnosis from fundus images. It also evaluates the performance of Vision Transformers (ViTs) relative to traditional CNN-based models. Methods: The proposed system uses three datasets: REFUGE, ORIGA, and G1020. It begins with YOLOv11 for object detection of the optic disc. Then, the optic disc (OD) and optic cup (OC) are segmented using U-Net with ResNet50, VGG16, and MobileNetV2 backbones, as well as MaskFormer with a Swin-Base backbone. Glaucoma is classified based on the vertical cup-to-disc ratio (vCDR). Results: MaskFormer outperforms all models in segmentation in all aspects, including IoU OD, IoU OC, DSC OD, and DSC OC, with scores of 88.29%, 91.09%, 93.83%, and 93.71%. For classification, it achieved accuracy and F1-scores of 84.03% and 84.56%. Conclusions: By relying on the interpretable features of the vCDR, the proposed framework enhances transparency and aligns well with the principles of explainable AI, thus offering a trustworthy solution for glaucoma screening. Our findings show that Vision Transformers offer a promising approach for achieving high segmentation performance with explainable, biomarker-driven diagnosis. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

20 pages, 18323 KB  
Article
MambaRA-GAN: Underwater Image Enhancement via Mamba and Intra-Domain Reconstruction Autoencoder
by Jiangyan Wu, Guanghui Zhang and Yugang Fan
J. Mar. Sci. Eng. 2025, 13(9), 1745; https://doi.org/10.3390/jmse13091745 - 10 Sep 2025
Abstract
Underwater images frequently suffer from severe quality degradation due to light attenuation and scattering effects, manifesting as color distortion, low contrast, and detail blurring. These issues significantly impair the performance of downstream tasks. Therefore, underwater image enhancement (UIE) becomes a key technology to [...] Read more.
Underwater images frequently suffer from severe quality degradation due to light attenuation and scattering effects, manifesting as color distortion, low contrast, and detail blurring. These issues significantly impair the performance of downstream tasks. Therefore, underwater image enhancement (UIE) becomes a key technology to solve underwater image degradation. However, existing data-driven UIE methods typically rely on difficult-to-acquire paired data for training, severely limiting their practical applicability. To overcome this limitation, this study proposes MambaRA-GAN, a novel unpaired UIE framework built upon a CycleGAN architecture, which introduces a novel integration of Mamba and intra-domain reconstruction autoencoders. The key innovations of our work are twofold: (1) We design a generator architecture based on a Triple-Gated Mamba (TG-Mamba) block. This design dynamically allocates feature channels to three parallel branches via learnable weights, achieving optimal fusion of CNN’s local feature extraction capabilities and Mamba’s global modeling capabilities. (2) We construct an intra-domain reconstruction autoencoder, isomorphic to the generator, to quantitatively assess the quality of reconstructed images within the cycle consistency loss. This introduces more effective structural information constraints during training. The experimental results demonstrate that the proposed method achieves significant improvements across five objective performance metrics. Visually, it effectively restores natural colors, enhances contrast, and preserves rich detail information, robustly validating its efficacy for the UIE task. Full article
(This article belongs to the Section Ocean Engineering)
21 pages, 29406 KB  
Article
Deep Learning-Based Contrail Segmentation in Thermal Infrared Satellite Cloud Images via Frequency-Domain Enhancement
by Shenhao Shi, Juncheng Wu, Kaixuan Yao and Qingxiang Meng
Remote Sens. 2025, 17(18), 3145; https://doi.org/10.3390/rs17183145 - 10 Sep 2025
Abstract
Aviation contrails significantly impact climate via radiative forcing, but their segmentation in thermal infrared satellite images is challenged by thin-layer structures, blurry edges, and cirrus cloud interference. We propose MFcontrail, a deep learning model integrating multi-axis attention and frequency-domain enhancement for precise contrail [...] Read more.
Aviation contrails significantly impact climate via radiative forcing, but their segmentation in thermal infrared satellite images is challenged by thin-layer structures, blurry edges, and cirrus cloud interference. We propose MFcontrail, a deep learning model integrating multi-axis attention and frequency-domain enhancement for precise contrail segmentation. It uses a MaxViT encoder to capture long-range spatial features, a FreqFusion decoder to preserve high-frequency edge details, and an edge-aware loss to refine boundary accuracy. Evaluations on OpenContrails and Landsat-8 datasets show that MFcontrail outperforms state-of-the-art methods: compared with DeepLabV3+, it achieves a 5.03% higher F1-score and 5.91% higher IoU on OpenContrails, with 3.43% F1-score and 4.07% IoU gains on Landsat-8. Ablation studies confirm the effectiveness of frequency-domain enhancement (contributing 69.4% of IoU improvement) and other key components. This work provides a high-precision tool for aviation climate research, highlighting frequency-domain strategies’ value in satellite cloud image analysis. Full article
43 pages, 2874 KB  
Article
Attention-Driven and Hierarchical Feature Fusion Network for Crop and Weed Segmentation with Fractal Dimension Estimation
by Rehan Akram, Jung Soo Kim, Min Su Jeong, Hafiz Ali Hamza Gondal, Muhammad Hamza Tariq, Muhammad Irfan and Kang Ryoung Park
Fractal Fract. 2025, 9(9), 592; https://doi.org/10.3390/fractalfract9090592 - 10 Sep 2025
Abstract
In precision agriculture, semantic segmentation enhances the crop yield by enabling precise disease monitoring, targeted herbicide application, and accurate crop–weed differentiation. This enhances yield; reduces the overuse of herbicides, water, and fertilizers; lowers labor costs; and promotes sustainable farming. Deep-learning-based methods are particularly [...] Read more.
In precision agriculture, semantic segmentation enhances the crop yield by enabling precise disease monitoring, targeted herbicide application, and accurate crop–weed differentiation. This enhances yield; reduces the overuse of herbicides, water, and fertilizers; lowers labor costs; and promotes sustainable farming. Deep-learning-based methods are particularly effective for crop and weed segmentation, and achieve potential results. Typically, segmentation is performed using homogeneous data (the same dataset is used for training and testing). However, previous studies, such as crop and weed segmentation in a heterogeneous data environment, using heterogeneous data (i.e., different datasets for training and testing) remain inaccurate. The proposed framework uses patch-based augmented limited training data within a heterogeneous environment to resolve the problems of degraded accuracy and the use of extensive data for training. We propose an attention-driven and hierarchical feature fusion network (AHFF-Net) comprising a flow-constrained convolutional block, hierarchical multi-stage fusion block, and attention-driven feature enhancement block. These blocks independently extract diverse fine-grained features and enhance the learning capabilities of the network. AHFF-Net is also combined with an open-source large language model (LLM)-based pesticide recommendation system made by large language model Meta AI (LLaMA). Additionally, a fractal dimension estimation method is incorporated into the system that provides valuable insights into the spatial distribution characteristics of crops and weeds. We conducted experiments using three publicly available datasets: BoniRob, Crop/Weed Field Image Dataset (CWFID), and Sunflower. For each experiment, we trained on one dataset and tested on another by reversing the process of the second experiment. The highest mean intersection of union (mIOU) of 65.3% and F1 score of 78.7% were achieved when training on the BoniRob dataset and testing on CWFID. This demonstrated that our method outperforms other state-of-the-art approaches. Full article
26 pages, 883 KB  
Review
Keratoconjunctivitis Sicca in Sjögren Disease: Diagnostic Challenges and Therapeutic Advances
by Muhammad Soyfoo, Elie Motulsky and Julie Sarrand
Int. J. Mol. Sci. 2025, 26(18), 8824; https://doi.org/10.3390/ijms26188824 - 10 Sep 2025
Abstract
Keratoconjunctivitis sicca (KCS), also commonly known as dry eye disease (DED), is one of the most prevalent and crippling features of Sjögren disease (SD), a chronic systemic autoimmune disorder featuring lymphocytic infiltration and progressive impairment of exocrine glands. KCS affects up to 95% [...] Read more.
Keratoconjunctivitis sicca (KCS), also commonly known as dry eye disease (DED), is one of the most prevalent and crippling features of Sjögren disease (SD), a chronic systemic autoimmune disorder featuring lymphocytic infiltration and progressive impairment of exocrine glands. KCS affects up to 95% of patients with SD and is often the earliest and most persistent manifestation, significantly compromising visual function, ocular comfort, and overall quality of life. Beyond the ocular surface, KCS mirrors a wider spectrum of immune dysregulation and epithelial damage characteristic of the disease, making it a valuable window into the underlying systemic pathology. The pathophysiology of KCS in SD is complex and multifactorial, involving an interplay between autoimmune-mediated lacrimal gland dysfunction, neuroimmune interactions, ocular surface inflammation, and epithelial instability. Tear film instability and epithelial injury result from the aberrant activation of innate and adaptive immunity, involving T and B lymphocytes, pro-inflammatory cytokines, and type I interferon pathways. Despite the clinical significance of KCS, its diagnosis remains challenging, with frequent discrepancies between subjective symptoms and objective findings. Traditional diagnostic tools often lack sensitivity and specificity, prompting the development of novel imaging techniques, tear film biomarkers, and standardized scoring systems. Concurrently, therapeutic strategies have evolved from palliative approaches to immunomodulatory and regenerative treatments, aiming to restore immune homeostasis and epithelial integrity. This review provides a comprehensive update on the pathogenesis, diagnostic landscape, and emerging treatments of KCS in the context of SD. Full article
(This article belongs to the Special Issue Molecular Advances in Dry Eye Syndrome)
Show Figures

Figure 1

24 pages, 7007 KB  
Article
M4MLF-YOLO: A Lightweight Semantic Segmentation Framework for Spacecraft Component Recognition
by Wenxin Yi, Zhang Zhang and Liang Chang
Remote Sens. 2025, 17(18), 3144; https://doi.org/10.3390/rs17183144 - 10 Sep 2025
Abstract
With the continuous advancement of on-orbit services and space intelligence sensing technologies, the efficient and accurate identification of spacecraft components has become increasingly critical. However, complex lighting conditions, background interference, and limited onboard computing resources present significant challenges to existing segmentation algorithms. To [...] Read more.
With the continuous advancement of on-orbit services and space intelligence sensing technologies, the efficient and accurate identification of spacecraft components has become increasingly critical. However, complex lighting conditions, background interference, and limited onboard computing resources present significant challenges to existing segmentation algorithms. To address these challenges, this paper proposes a lightweight spacecraft component segmentation framework for on-orbit applications, termed M4MLF-YOLO. Based on the YOLOv5 architecture, we propose a refined lightweight design strategy that aims to balance segmentation accuracy and resource consumption in satellite-based scenarios. MobileNetV4 is adopted as the backbone network to minimize computational overhead. Additionally, a Multi-Scale Fourier Adaptive Calibration Module (MFAC) is designed to enhance multi-scale feature modeling and boundary discrimination capabilities in the frequency domain. We also introduce a Linear Deformable Convolution (LDConv) to explicitly control the spatial sampling span and distribution of the convolution kernel, thereby linearly adjusting the receptive field coverage range to improve feature extraction capabilities while effectively reducing computational costs. Furthermore, the efficient C3-Faster module is integrated to enhance channel interaction and feature fusion efficiency. A high-quality spacecraft image dataset, comprising both real and synthetic images, was constructed, covering various backgrounds and component types, including solar panels, antennas, payload instruments, thrusters, and optical payloads. Environment-aware preprocessing and enhancement strategies were applied to improve model robustness. Experimental results demonstrate that M4MLF-YOLO achieves excellent segmentation performance while maintaining low model complexity, with precision reaching 95.1% and recall reaching 88.3%, representing improvements of 1.9% and 3.9% over YOLOv5s, respectively. The mAP@0.5 also reached 93.4%. In terms of lightweight design, the model parameter count and computational complexity were reduced by 36.5% and 24.6%, respectively. These results validate that the proposed method significantly enhances deployment efficiency while preserving segmentation accuracy, showcasing promising potential for satellite-based visual perception applications. Full article
Show Figures

Figure 1

15 pages, 1093 KB  
Article
A Multimodal Power Sample Feature Migration Method Based on Dual Cross-Modal Information Decoupling
by Zhenyu Chen, Huaguang Yan, Jianguang Du, Yuhao Zhou, Yi Chen, Yunfeng Yan and Shuai Zhao
Appl. Sci. 2025, 15(18), 9913; https://doi.org/10.3390/app15189913 - 10 Sep 2025
Abstract
With the rapid development of energy transition and power system informatization, the efficient integration and feature migration of multimodal power data have become critical challenges for intelligent power systems. Existing methods often overlook fine-grained semantic relationships in cross-modal alignment, leading to low information [...] Read more.
With the rapid development of energy transition and power system informatization, the efficient integration and feature migration of multimodal power data have become critical challenges for intelligent power systems. Existing methods often overlook fine-grained semantic relationships in cross-modal alignment, leading to low information utilization. This paper proposes a multimodal power sample feature migration method based on dual cross-modal information decoupling. By introducing a fine-grained image–text alignment strategy and a dual-stream attention mechanism, deep integration and efficient migration of multimodal features are achieved. Experiments demonstrate that the proposed method outperforms baseline models (e.g., LLaVA, Qwen) in power scenario description (CSD), event localization (CELC), and knowledge question answering (CKQ), with significant improvements of up to 12.8% in key metrics such as image captioning (IC) and grounded captioning (GC). The method provides a robust solution for multimodal feature migration in power inspection and real-time monitoring, showing high practical value in industrial applications. Full article
Show Figures

Figure 1

29 pages, 5334 KB  
Article
A Novel Self-Recovery Fragile Watermarking Scheme Based on Convolutional Autoencoder
by Chin-Feng Lee, Tong-Ming Li, Iuon-Chang Lin and Anis Ur Rehman
Electronics 2025, 14(18), 3595; https://doi.org/10.3390/electronics14183595 - 10 Sep 2025
Abstract
In the digital era where images are easily accessible, concerns about image authenticity and integrity are increasing. To address this, we propose a deep learning-based fragile watermarking method for secure image authentication and content recovery. The method utilizes bottleneck features extracted by the [...] Read more.
In the digital era where images are easily accessible, concerns about image authenticity and integrity are increasing. To address this, we propose a deep learning-based fragile watermarking method for secure image authentication and content recovery. The method utilizes bottleneck features extracted by the convolutional encoder to carry both authentication and recovery information and employs deconvolution at the decoder to reconstruct image content. Additionally, the Arnold Transform is applied to scramble feature information, effectively enhancing resistance to collage attacks. At the detection stage, block voting and morphological closing operations improve tamper localization accuracy and robustness. Experiments tested various tampering ratios, with performance evaluated by PSNR, SSIM, precision, recall, and F1-score. Experiments under varying tampering ratios demonstrate that the proposed method maintains high visual quality and achieves reliable tamper detection and recovery, even at 75% tampering. Evaluation metrics including PSNR, SSIM, precision, recall, and F1-score confirm the effectiveness and practical applicability of the method. Full article
(This article belongs to the Special Issue Digital Signal and Image Processing for Multimedia Technology)
Show Figures

Figure 1

23 pages, 5348 KB  
Article
A Symmetry-Aware Multi-Attention Framework for Bird Nest Detection on Railway Catenary Systems
by Peiting Shan, Wei Feng, Shuntian Lou, Gabriel Dauphin and Wenxing Bao
Symmetry 2025, 17(9), 1505; https://doi.org/10.3390/sym17091505 - 10 Sep 2025
Abstract
Railway service interruptions and electrical hazards often arise due to bird nests concealed within the intricate, highly symmetric overhead catenary networks of high-speed lines. These nests are difficult to pinpoint automatically, not only because they are diminutive and often merge visually with the [...] Read more.
Railway service interruptions and electrical hazards often arise due to bird nests concealed within the intricate, highly symmetric overhead catenary networks of high-speed lines. These nests are difficult to pinpoint automatically, not only because they are diminutive and often merge visually with the surroundings but also due to occlusions and the persistent lack of substantial labeled datasets. To address this bottleneck, this work presents the High-Speed Railway Catenary Nest Dataset (HRC-Nest), merging 800 authentic images and 1000 synthetic samples to capture a spectrum of scenarios. Building on the symmetry of catenary structures—where nests appear as localized asymmetries—the Symmetry-Aware Railway Nest Detection Framework (RNDF) is proposed, an enhanced YOLOv12 system for accurate and robust nest detection in symmetric high-speed railway catenary environments. With the A2C2f_HRAMi design, the RNDF learns from multi-level features by unifying residual and hierarchical attention strategies. The SCSA component boosts the recognition in visually cluttered or obstructed settings further by jointly processing spatial and channel-wise signals. To sharpen the detection accuracy, particularly for subtle, hidden nests, the Focaler-GIoU loss guides bounding box optimization. Comparative studies show that the RNDF consistently outperforms recent detectors, surpassing the YOLOv12 baseline by 5.95% mAP@0.5 and 26.16% mAP@0.5:0.95, underscoring its suitability for symmetry-aware, real-world catenary anomaly monitoring. Full article
(This article belongs to the Special Issue Symmetry/Asymmetry in Digital Image Processing)
Show Figures

Figure 1

19 pages, 2435 KB  
Article
Image Sensor-Supported Multimodal Attention Modeling for Educational Intelligence
by Yanlin Chen, Yingqiu Yang, Zeyu Lan, Xinyuan Chen, Haoyuan Zhan, Lingxi Yu and Yan Zhan
Sensors 2025, 25(18), 5640; https://doi.org/10.3390/s25185640 - 10 Sep 2025
Abstract
To address the limitations of low fusion efficiency and insufficient personalization in multimodal perception for educational intelligence, a novel deep learning framework is proposed that integrates image sensor data with textual and contextual information through a cross-modal attention mechanism. The architecture employs a [...] Read more.
To address the limitations of low fusion efficiency and insufficient personalization in multimodal perception for educational intelligence, a novel deep learning framework is proposed that integrates image sensor data with textual and contextual information through a cross-modal attention mechanism. The architecture employs a cross-modal alignment module to achieve fine-grained semantic correspondence between visual features captured by image sensors and associated textual elements, followed by a personalized feedback generator that incorporates learner background and task context embeddings to produce adaptive educational guidance. A cognitive weakness highlighter is introduced to enhance the discriminability of task-relevant features, enabling explicit localization and interpretation of conceptual gaps. Experiments show the proposed method outperforms conventional fusion and unimodal baselines with 92.37% accuracy, 91.28% recall, and 90.84% precision. Cross-task and noise-robustness tests confirm its stability, while ablation studies highlight the fusion module’s +4.2% accuracy gain and the attention mechanism’s +3.8% recall and +3.5% precision improvements. These results establish the proposed method as a transferable, high-performance solution for next-generation adaptive learning systems, offering precise, explainable, and context-aware feedback grounded in advanced multimodal perception modeling. Full article
Show Figures

Figure 1

27 pages, 13123 KB  
Article
Symmetric Boundary-Enhanced U-Net with Mamba Architecture for Glomerular Segmentation in Renal Pathological Images
by Shengnan Zhang, Xinming Cui, Guangkun Ma and Ronghui Tian
Symmetry 2025, 17(9), 1506; https://doi.org/10.3390/sym17091506 - 10 Sep 2025
Abstract
Accurate glomerular segmentation in renal pathological images is a key challenge for chronic kidney disease diagnosis and assessment. Due to the high visual similarity between pathological glomeruli and surrounding tissues in color, texture, and morphology, significant “camouflage phenomena” exist, leading to boundary identification [...] Read more.
Accurate glomerular segmentation in renal pathological images is a key challenge for chronic kidney disease diagnosis and assessment. Due to the high visual similarity between pathological glomeruli and surrounding tissues in color, texture, and morphology, significant “camouflage phenomena” exist, leading to boundary identification difficulties. To address this problem, we propose BM-UNet, a novel segmentation framework that embeds boundary guidance mechanisms into a Mamba architecture with a symmetric encoder–decoder design. The framework enhances feature transmission through explicit boundary detection, incorporating four core modules designed for key challenges in pathological image segmentation. The Multi-scale Adaptive Fusion (MAF) module processes irregular tissue morphology, the Hybrid Boundary Detection (HBD) module handles boundary feature extraction, the Boundary-guided Attention (BGA) module achieves boundary-aware feature refinement, and the Mamba-based Fused Decoder Block (MFDB) completes boundary-preserving reconstruction. By introducing explicit boundary supervision mechanisms, the framework achieves significant segmentation accuracy improvements while maintaining linear computational complexity. Validation on the KPIs2024 glomerular dataset and HuBMAP renal tissue samples demonstrates that BM-UNet achieves a 92.4–95.3% mean Intersection over Union across different CKD pathological conditions, with a 4.57% improvement over the Mamba baseline and a processing speed of 113.7 FPS. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

18 pages, 808 KB  
Article
Towards AI-Based Strep Throat Detection and Interpretation for Remote Australian Indigenous Communities
by Prasanna Asokan, Thanh Thu Truong, Duc Son Pham, Kit Yan Chan, Susannah Soon, Andrew Maiorana and Cate Hollingsworth
Sensors 2025, 25(18), 5636; https://doi.org/10.3390/s25185636 - 10 Sep 2025
Abstract
Streptococcus pharyngitis (strep throat) poses a significant health challenge in rural and remote Indigenous communities in Australia, where access to medical resources is limited. Delays in diagnosis and treatment increase the risk of serious complications, including acute rheumatic fever and rheumatic heart disease. [...] Read more.
Streptococcus pharyngitis (strep throat) poses a significant health challenge in rural and remote Indigenous communities in Australia, where access to medical resources is limited. Delays in diagnosis and treatment increase the risk of serious complications, including acute rheumatic fever and rheumatic heart disease. This paper presents a proof-of-concept AI-based diagnostic model designed to support clinicians in underserved communities. The model combines a lightweight Swin Transformer–based image classifier with a BLIP-2-based explainable image annotation system. The classifier predicts strep throat from throat images, while the explainable model enhances transparency by identifying key clinical features such as tonsillar swelling, erythema, and exudate, with synthetic labels generated using GPT-4o-mini. The classifier achieves 97.1% accuracy and an ROC-AUC of 0.993, with an inference time of 13.8 ms and a model size of 28 million parameters; these results demonstrate suitability for deployment in resource-constrained settings. As a proof-of-concept, this work illustrates the potential of AI-assisted diagnostics to improve healthcare access and could benefit similar research efforts that support clinical decision-making in remote and underserved regions. Full article
Show Figures

Figure 1

32 pages, 8264 KB  
Article
SATRNet: Self-Attention-Aided Deep Unfolding Tensor Representation Network for Robust Hyperspectral Anomaly Detection
by Jing Yang, Jianbin Zhao, Lu Chen, Haorui Ning and Ying Li
Remote Sens. 2025, 17(18), 3137; https://doi.org/10.3390/rs17183137 - 10 Sep 2025
Abstract
Hyperspectral anomaly detection (HAD) aims to separate subtle anomalies of a given hyperspectral image (HSI) from its background, which is a hot topic as well as a challenging inverse problem. Despite the significant success of the deep learning-based HAD methods, they are hard [...] Read more.
Hyperspectral anomaly detection (HAD) aims to separate subtle anomalies of a given hyperspectral image (HSI) from its background, which is a hot topic as well as a challenging inverse problem. Despite the significant success of the deep learning-based HAD methods, they are hard to interpret due to their black-box nature. Meanwhile, deep learning methods suffer from the identity mapping (IM) problem, referring to the network excessively focusing on the precise reconstruction of the background while neglecting the appropriate representation of anomalies. To this end, this paper proposes a self-attention-aided deep unfolding tensor representation network (SATRNet) for interpretable HAD by solving the tensor representation (TR)-based optimization model within the framework of deep networks. In particular, a Self-Attention Learning Module (SALM) was first designed to extract discriminative features of the input HSI. The HAD problem was then formulated as a tensor representation problem by exploring both the low-rankness of the background and the sparsity of the anomaly. A Weight Learning Module (WLM) exploring local details was also generated for precise background reconstruction. Finally, a deep network was built to solve the TR-based problem through unfolding and parameterizing the iterative optimization algorithm. The proposed SATRNet prevents the network from learning meaningless mappings, making the network interpretable to some extent while essentially solving the IM problem. The effectiveness of the proposed SATRNet is validated on 11 benchmark HSI datasets. Notably, the performance of SATRNet against adversarial attacks is also investigated in the experimentation, which is the first work exploring adversarial robustness in HAD to the best of our knowledge. Full article
Show Figures

Figure 1

Back to TopTop