Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,124)

Search Parameters:
Keywords = semi-supervised learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 1910 KB  
Article
Semi-Supervised Generative Adversarial Networks (GANs) for Adhesion Condition Identification in Intelligent and Autonomous Railway Systems
by Sanaullah Mehran, Khakoo Mal, Imtiaz Hussain, Dileep Kumar, Tarique Rafique Memon and Tayab Din Memon
AI 2026, 7(2), 78; https://doi.org/10.3390/ai7020078 - 18 Feb 2026
Abstract
Safe and reliable railway operation forms an integral part of autonomous transport systems and depends on accurate knowledge of the adhesion conditions. Both the underestimation and overestimation of adhesion can compromise real-time decision-making in traction and braking control, leading to accidents or excessive [...] Read more.
Safe and reliable railway operation forms an integral part of autonomous transport systems and depends on accurate knowledge of the adhesion conditions. Both the underestimation and overestimation of adhesion can compromise real-time decision-making in traction and braking control, leading to accidents or excessive wear at the wheel–rail interface. Although limited research has explored the estimation of adhesion forces using data-driven algorithms, most existing approaches lack self-reliance and fail to adequately capture low adhesion levels, which are critical to identify. Moreover, obtaining labelled experimental data remains a significant challenge in adopting data-driven solutions for domain-specific problems. This study implements self-reliant deep learning (DL) models as perception modules for intelligent railway systems, enabling low adhesion identification by training on raw time sequences. In the second phase, to address the challenge of label acquisition, a semi-supervised generative adversarial network (SGAN) is developed. Compared to the supervised algorithms, the SGAN achieved superior performance, with 98.38% accuracy, 98.42% precision, and 98.28% F1-score in identifying seven different adhesion conditions. In contrast, the MLP and 1D-CNN models achieved accuracy of 91% and 93.88%, respectively. These findings demonstrate the potential of SGAN-based data-driven perception for enhancing autonomy, adaptability, and fault diagnosis in intelligent rail and robotic mobility systems. The proposed approach offers an efficient and scalable solution for real-time railway condition monitoring and fault identification, eliminating the overhead associated with manual data labelling. Full article
(This article belongs to the Special Issue Development and Design of Autonomous Robot)
22 pages, 11711 KB  
Article
Mitigating Urban Flooding Through Residential Rainwater Harvesting Using GIS and HEC-HMS
by Isabel Lopez and Ivonne Santiago
Water 2026, 18(4), 487; https://doi.org/10.3390/w18040487 - 14 Feb 2026
Viewed by 117
Abstract
As urbanization expands, the loss of pervious surfaces has led to greater stormwater runoff and contributed to an increase in urban flooding—localized flooding in areas not formally designated as flood zones. This study evaluates the potential of decentralized active rainwater harvesting (RWH) to [...] Read more.
As urbanization expands, the loss of pervious surfaces has led to greater stormwater runoff and contributed to an increase in urban flooding—localized flooding in areas not formally designated as flood zones. This study evaluates the potential of decentralized active rainwater harvesting (RWH) to mitigate urban flooding in semi-arid urban environments. A neighborhood in northeast El Paso, Texas, was selected as a pilot site. Using a GIS-HEC-HMS modeling framework, approximately 9000 residential parcels were analyzed to assess rooftop harvesting capacity, runoff potential, and system feasibility under different adoption rates and antecedent moisture conditions. Land cover and building footprints were extracted using supervised machine learning to generate stormwater runoff parameters and catchment areas for rainfall-runoff simulations for storms with return periods ranging from 1 to 50 years. The results indicate that for 1- and 2-year storms, a 25% adoption rate may reduce street runoff by 16–19% from 13.1 to 10.6 × 103 m3 and from 31 to 26.1 × 103 m3. Increasing adoption to 50% yields substantially greater reductions of approximately 30–36%. Even higher-magnitude storms (5- and 10-year events) experience measurable decreases in runoff volume, with reductions of 10% for the 5-year storms and up to 10.4% for the 10-year storm at the 25% adoption and 20–22% across the same events at 50% adoption. Overall, the results of this study demonstrate that GIS and HEC-HMS are effective tools for evaluating urban flood mitigation strategies, and that decentralized RWH offers a viable method for reducing flood risk in urbanized settings when adoption levels and storage capacities are considered. Full article
(This article belongs to the Section Urban Water Management)
Show Figures

Figure 1

32 pages, 5302 KB  
Article
Class-Driven Robust Non-Negative Matrix Factorization with Dual-Hypergraph Regularization for Data Clustering
by Haiyan Gao and Gaigai Zhou
Symmetry 2026, 18(2), 351; https://doi.org/10.3390/sym18020351 - 13 Feb 2026
Viewed by 69
Abstract
Traditional non-negative matrix factorization (NMF) faces challenges when dealing with complex data, primarily characterized by sensitivity to noise, neglect of data geometric structure, and inability to effectively utilize supervised information. To address these limitations, this paper proposes a class-driven robust non-negative matrix factorization [...] Read more.
Traditional non-negative matrix factorization (NMF) faces challenges when dealing with complex data, primarily characterized by sensitivity to noise, neglect of data geometric structure, and inability to effectively utilize supervised information. To address these limitations, this paper proposes a class-driven robust non-negative matrix factorization with dual-hypergraph regularization (CRNMFDH) framework. The core contributions of this framework include the following: Firstly, the design of a novel dual-hypergraph regularization term that symmetrically captures and preserves the higher-order geometric structures of both the sample space and feature space, establishing a mutually reinforcing topological relationship between them. Secondly, an introduction of a class-driven mechanism to effectively integrate label information into the decomposition process, significantly enhancing the discriminative capability of the low-dimensional representations. Finally, the employment of a loss function based on correntropy to replace the traditional Euclidean distance, thereby enhancing the model’s robustness against noise and outliers. Extensive experiments across nine datasets demonstrate that CRNMFDH significantly outperforms existing state-of-the-art algorithms in multiple clustering evaluation metrics and noise robustness, providing an effective new solution for complex data clustering tasks. Full article
(This article belongs to the Section Computer)
24 pages, 4132 KB  
Article
Unsupervised Learning Framework for Cyber Threat Detection, Anomaly Identification, and Alert Prioritization
by Emmanuel Okafor and Seokhee Lee
Appl. Sci. 2026, 16(4), 1884; https://doi.org/10.3390/app16041884 - 13 Feb 2026
Viewed by 150
Abstract
Conventional Security Operations Center (SOC) solutions struggle to process representative operational alert streams efficiently and adapt to evolving cyber threats, highlighting the need for automated, intelligent threat detection and prioritization. This study presents a custom AI-driven framework that leverages unsupervised learning techniques to [...] Read more.
Conventional Security Operations Center (SOC) solutions struggle to process representative operational alert streams efficiently and adapt to evolving cyber threats, highlighting the need for automated, intelligent threat detection and prioritization. This study presents a custom AI-driven framework that leverages unsupervised learning techniques to support SOC analysts in cyber threat detection, anomaly identification, and alert prioritization. The framework applies several clustering methods: HDBSCAN, DBSCAN, KMeans, and Gaussian Mixture Models for alert segmentation, and integrates anomaly detection using LOF and Isolation Forest, complemented by semi-supervised detection via One-Class SVM. Using textual, categorical, and numerical features from Wazuh alerts across three datasets, the system performs clustering and anomaly detection in the original high-dimensional feature space, with UMAP applied solely for two-dimensional visualization. HDBSCAN consistently produced well-separated clusters with effective noise detection, while, Isolation Forest evaluated via 10-fold cross-validation exhibited stable anomaly flagging and clear score separation across both cyber alert event data and synthetic threat injection experiments. Furthermore, the framework formulates a composite priority ranking that integrates anomaly severity, cluster rarity, and SOC contextual weighting, yielding actionable alert rankings. An interactive, analyst-centric dashboard enables SOC teams to explore top alerts, clusters, associated MITRE techniques, priority rankings, and geolocation data, providing insights while preserving human oversight. Overall, the proposed system transforms complex alert streams into structured insights, enhancing SOC situational awareness, decision support, and operational efficiency. Full article
Show Figures

Figure 1

28 pages, 1177 KB  
Article
Context-Aware Code Review Automation: A Retrieval-Augmented Approach
by Büşra İçöz and Göksel Biricik
Appl. Sci. 2026, 16(4), 1875; https://doi.org/10.3390/app16041875 - 13 Feb 2026
Viewed by 157
Abstract
Manual code review is essential for software quality, but often slows down development cycles due to the high time demands on developers. In this study, we propose an automated solution for Python (version 3.13) projects that generates code review comments by combining Large [...] Read more.
Manual code review is essential for software quality, but often slows down development cycles due to the high time demands on developers. In this study, we propose an automated solution for Python (version 3.13) projects that generates code review comments by combining Large Language Models (LLMs) with Retrieval-Augmented Generation (RAG). To achieve this, we first curated a dataset from GitHub pull requests (PRs) using the GitHub REST Application Programming Interface (API) (version 2022-11-28) and classified comments into semantic categories using a semi-supervised Support Vector Machine (SVM) model. During the review process, our system uses a vector database to retrieve the top-k most relevant historical comments, providing context for a diverse spectrum of open-weights LLMs, including DeepSeek-Coder-33B, Qwen2.5-Coder-32B, Codestral-22B, CodeLlama-13B, Mistral-Instruct-7B, and Phi-3-Mini. We evaluated the system using a multi-step validation that combined standard metrics (BLEU-4, ROUGE-L, cosine similarity) with an LLM-as-a-Judge approach, and verified the results through targeted human review to ensure consistency with expert standards. The findings show that retrieval augmentation improves feedback relevance for larger models, with DeepSeek-Coder’s alignment score increasing by 17.9% at a retrieval depth of k = 3. In contrast, smaller models such as Phi-3-Mini suffered from context collapse, where too much context reduced accuracy. To manage this trade-off, we built a hybrid expert system that routes each task to the most suitable model. Our results indicate that the proposed approach improved performance by 13.2% compared to the zero-shot baseline (k = 0). In addition, our proposed system reduces hallucinations and generates comments that closely align with the standards expected from the experts. Full article
(This article belongs to the Special Issue Artificial Intelligence in Software Engineering)
Show Figures

Figure 1

26 pages, 5988 KB  
Article
Limited-Annotation Seed Segmentation for Analyzing the Impact of Unsound Corn on Storage Quality
by Kuibin Zhao, Lei Lu, Hongyi Ge, Pengtao Lv and Jinpei Li
Agriculture 2026, 16(4), 421; https://doi.org/10.3390/agriculture16040421 - 12 Feb 2026
Viewed by 105
Abstract
Grain quality inspection is crucial for seed stored, with image segmentation playing a key role in this process. However, existing methods face challenges such as high computational costs, expensive data annotation, and data privacy concerns, which hinder the acquisition of large-scale labeled datasets [...] Read more.
Grain quality inspection is crucial for seed stored, with image segmentation playing a key role in this process. However, existing methods face challenges such as high computational costs, expensive data annotation, and data privacy concerns, which hinder the acquisition of large-scale labeled datasets and limit model performance. To overcome these challenges, we propose a novel semi-supervised learning (SSL) paradigm for seed segmentation. Our approach integrates VMUNet and UNet into a unified framework, combining UNet’s capacity for fine-grained detail extraction with VMUNet’s strengths in global semantic model, enabling richer pixel-level feature representation. We introduce an orthogonal attention mechanism into the VMUNet encoder to model feature dependencies across channel, spatial, and scale dimensions, improving information fusion and feature enhancement. Additionally, a perturbation strategy is applied in the dual-branch decoder, combined with consistency regularization, to enhance robustness and generalization. This helps mitigate overfitting and reduces excessive reliance on boundary details during decoding. Experimental results on a corn seed dataset show that the proposed method achieves 91.2% accuracy with 100% labeled data and 91.9% with only 50% labeled data, outperforming fully supervised methods by 0.6%. These results demonstrate the method’s high segmentation performance and practical potential while maintaining data privacy. These results confirm that OAMamba provides an accurate, robust, and annotation-efficient solution for corn seed segmentation, showing strong potential for practical deployment in agricultural intelligent inspection systems. Full article
(This article belongs to the Section Agricultural Product Quality and Safety)
Show Figures

Figure 1

10 pages, 12951 KB  
Proceeding Paper
A Forest Mapping Model for Algeria Using Noisy Labels and Few Clean Data
by Lilia Ammar Khodja, Meziane Iftene and Mohammed El Amin Larabi
Eng. Proc. 2026, 124(1), 19; https://doi.org/10.3390/engproc2026124019 - 6 Feb 2026
Viewed by 155
Abstract
This study proposes a forest mapping framework for Algeria that addresses the challenge of limited clean data and noisy global land cover labels. The approach combines a small set of manually curated annotations with noisy ESA WorldCover data, leveraging Sentinel-2 multispectral imagery and [...] Read more.
This study proposes a forest mapping framework for Algeria that addresses the challenge of limited clean data and noisy global land cover labels. The approach combines a small set of manually curated annotations with noisy ESA WorldCover data, leveraging Sentinel-2 multispectral imagery and Digital Elevation Model (DEM) features such as slope, aspect, and the Normalized Difference Vegetation Index (NDVI). A modified ResNet-18 architecture was fine-tuned using both clean and pseudo-labeled noisy data, enabling the model to effectively mitigate label noise. The framework achieved an overall accuracy of 98.5%, demonstrating strong generalization across Algeria’s diverse forest ecosystems. These results highlight the potential of semi-supervised deep learning to improve large-scale forest monitoring, with applications in conservation, sustainable resource management, and climate change mitigation. Full article
(This article belongs to the Proceedings of The 6th International Electronic Conference on Applied Sciences)
Show Figures

Figure 1

23 pages, 15010 KB  
Article
Hybrid Mamba–Graph Fusion with Multi-Stage Pseudo-Label Refinement for Semi-Supervised Hyperspectral–LiDAR Classification
by Khanzada Muzammil Hussain, Keyun Zhao, Sachal Perviaz and Ying Li
Sensors 2026, 26(3), 1005; https://doi.org/10.3390/s26031005 - 3 Feb 2026
Viewed by 340
Abstract
Semi-supervised joint classification of Hyperspectral Images (HSIs) and LiDAR-derived Digital Surface Models (DSMs) remains challenging due to scarcity of labeled pixels, strong intra-class variability, and the heterogeneous nature of spectral and elevation features. In this work, we propose a Hybrid Mamba–Graph Fusion Network [...] Read more.
Semi-supervised joint classification of Hyperspectral Images (HSIs) and LiDAR-derived Digital Surface Models (DSMs) remains challenging due to scarcity of labeled pixels, strong intra-class variability, and the heterogeneous nature of spectral and elevation features. In this work, we propose a Hybrid Mamba–Graph Fusion Network (HMGF-Net) with Multi-Stage Pseudo-Label Refinement (MS-PLR) for semi-supervised hyperspectral–LiDAR classification. The framework employs a spectral–spatial HSI backbone combining 3D–2D convolutions, a compact LiDAR CNN encoder, Mamba-style state-space sequence blocks for long-range spectral and cross-modal dependency modeling, and a graph fusion module that propagates information over a heterogeneous pixel graph. Semi-supervised learning is realized via a three-stage pseudolabeling pipeline that progressively filters, smooths, and re-weights pseudolabels based on prediction confidence, spatial–spectral consistency, and graph neighborhood agreement. We validate HMGF-Net on three benchmark hyperspectral–LiDAR datasets. Compared with a set of eight state-of-the-art (SOTA) baselines, including 3D-CNNs, SSRN, HybridSN, transformer-based models such as SpectralFormer, multimodal CNN–GCN fusion networks, and recent semi-supervised methods, the proposed approach delivers consistent gains in overall accuracy, average accuracy, and Cohen’s kappa, especially in low-label regimes (10% labeled pixels). The results highlight that the synergy between sequence modeling and graph reasoning in combination with carefully designed pseudolabel refinement is essential to maximizing the benefit of abundant unlabeled samples in multimodal remote sensing scenarios. Full article
(This article belongs to the Special Issue Progress in LiDAR Technologies and Applications)
Show Figures

Figure 1

12 pages, 1209 KB  
Article
Deep Learning-Based Semantic Segmentation and Classification of Otoscopic Images for Otitis Media Diagnosis and Health Promotion
by Chien-Yi Yang, Che-Jui Lee, Wen-Sen Lai, Kuan-Yu Chen, Chung-Feng Kuo, Chieh Hsing Liu and Shao-Cheng Liu
Diagnostics 2026, 16(3), 467; https://doi.org/10.3390/diagnostics16030467 - 2 Feb 2026
Viewed by 333
Abstract
Background/Objectives: Otitis media (OM), including acute otitis media (AOM) and chronic otitis media (COM), is a common middle ear disease that can lead to significant morbidity if not accurately diagnosed. Otoscopic interpretation remains subjective and operator-dependent, underscoring the need for objective and reproducible [...] Read more.
Background/Objectives: Otitis media (OM), including acute otitis media (AOM) and chronic otitis media (COM), is a common middle ear disease that can lead to significant morbidity if not accurately diagnosed. Otoscopic interpretation remains subjective and operator-dependent, underscoring the need for objective and reproducible diagnostic support. Recent advances in artificial intelligence (AI) offer promising solutions for automated otoscopic image analysis. Methods: We developed an AI-based diagnostic framework consisting of three sequential steps: (1) semi-supervised learning for automatic recognition and semantic segmentation of tympanic membrane structures, (2) region-based feature extraction, and (3) disease classification. A total of 607 clinical otoscopic images were retrospectively collected, including normal ears (n = 220), AOM (n = 157), and COM with tympanic membrane perforation (n = 230). Among these, 485 images were used for training and 122 for independent testing. Semantic segmentation of five anatomically relevant regions was performed using multiple convolutional neural network architectures, including U-Net, PSPNet, HRNet, and DeepLabV3+. Following segmentation, color and texture features were extracted from each region and used to train a neural network-based classifier to differentiate disease states. Results: Among the evaluated segmentation models, U-Net demonstrated superior performance, achieving an overall pixel accuracy of 96.76% and a mean Dice similarity coefficient of 71.68%. The segmented regions enabled reliable extraction of discriminative chromatic and texture features. In the final classification stage, the proposed framework achieved diagnostic accuracies of 100% for normal ears, 100% for AOM, and 91.3% for COM on the independent test set, with an overall accuracy of 96.72%. Conclusions: This study demonstrates that a semi-supervised, segmentation-driven AI pipeline integrating feature extraction and classification can achieve high diagnostic accuracy for otitis media. The proposed framework offers a clinically interpretable and fully automated approach that may enhance diagnostic consistency, support clinical decision-making, and facilitate scalable otoscopic assessment in diverse healthcare screening settings for disease prevention and health education. Full article
(This article belongs to the Special Issue AI-Assisted Diagnostics in Telemedicine and Digital Health)
Show Figures

Figure 1

34 pages, 3680 KB  
Article
A Semi-Supervised Transformer with a Curriculum Training Pipeline for Remote Sensing Image Semantic Segmentation
by Peizhuo Liu, Hongbo Zhu, Xiaofei Mi, Yuke Meng, Huijie Zhao and Xingfa Gu
Remote Sens. 2026, 18(3), 480; https://doi.org/10.3390/rs18030480 - 2 Feb 2026
Viewed by 210
Abstract
Semantic segmentation of remote sensing images is crucial for geospatial applications but is severely hampered by the prohibitive cost of pixel-level annotations. Although semi-supervised learning (SSL) offers a solution by leveraging unlabeled data, its application to Vision Transformers (ViTs) often encounters overfitting and [...] Read more.
Semantic segmentation of remote sensing images is crucial for geospatial applications but is severely hampered by the prohibitive cost of pixel-level annotations. Although semi-supervised learning (SSL) offers a solution by leveraging unlabeled data, its application to Vision Transformers (ViTs) often encounters overfitting and even training instability under extreme label scarcity. To tackle these challenges, we propose a Curriculum-based Self-supervised and Semi-supervised Pipeline (CSSP). The pipeline adopts a staged, easy-to-hard training strategy, commencing with in-domain pretraining for robust feature representation, followed by a carefully designed finetuning stage to prevent overfitting. The pipeline further integrates a novel Difficulty-Adaptive ClassMix (DA-ClassMix) augmentation that dynamically reinforces underperforming categories and a Progressive Intensity Adaptation (PIA) strategy that systematically escalates augmentation strength to maximize model generalization. Extensive evaluations on the Potsdam, Vaihingen, and Inria datasets demonstrate state-of-the-art performance. Notably, with only 1/32 of the labeled data on the Potsdam dataset, the CSSP reaches 82.16% mIoU, nearly matching the fully supervised result (82.24%). Furthermore, we extend the CSSP to a semi-supervised domain adaptation (SSDA) scenario, termed Cross-Domain CSSP (CDCSSP), which outperforms existing SSDA and unsupervised domain adaptation (UDA) methods. This work establishes a stable and highly effective framework for training ViT-based segmentation models with minimal annotation overhead. Full article
Show Figures

Figure 1

24 pages, 27057 KB  
Article
Super-Resolution Reconstruction of Gravity Data Using Semi-Supervised Dual Regression Learning
by Bode Jia, Jian Sun, Xiangfeng Geng, Xiaolei Wan and Huaishan Liu
Remote Sens. 2026, 18(3), 453; https://doi.org/10.3390/rs18030453 - 1 Feb 2026
Viewed by 291
Abstract
High-resolution (HR) marine gravity data are critical for geophysical modeling, seafloor mapping, and tectonic analysis. However, acquiring such data remains challenging due to the inherent trade-offs between distinct measurement sources. While shipborne gravity surveys offer high accuracy and resolution, they are spatially sparse [...] Read more.
High-resolution (HR) marine gravity data are critical for geophysical modeling, seafloor mapping, and tectonic analysis. However, acquiring such data remains challenging due to the inherent trade-offs between distinct measurement sources. While shipborne gravity surveys offer high accuracy and resolution, they are spatially sparse and geographically restricted; conversely, satellite altimetry provides global coverage but comes at the expense of reduced resolution and increased noise. To address this challenge, we propose a semi-supervised dual regression learning (SDRL) framework for gravity field super-resolution (SR) that synergizes the strengths of both data types. By jointly training on a limited number of paired shipborne-satellite samples and a large set of unpaired satellite observations, SDRL leverages cycle-consistent learning to preserve cross-domain structural integrity and enhance generalization. Extensive experiments under varying data conditions—including noisy, ideal, and label-scarce scenarios—demonstrate that SDRL consistently outperforms purely supervised models in terms of structural similarity and error reduction. Moreover, SDRL exhibits strong robustness against data imperfections and generalizes effectively to geophysically distinct test regions. These results highlight the practical advantages of semi-supervised learning for global marine gravity field reconstruction, particularly in real-world settings where high-quality labeled data are scarce. Full article
(This article belongs to the Special Issue Advances in Multi-Source Remote Sensing Data Fusion and Analysis)
Show Figures

Figure 1

24 pages, 3086 KB  
Article
Semi-Supervised Hyperspectral Reconstruction from RGB Images via Spectrally Aware Mini-Patch Calibration
by Runmu Su, Haosong Huang, Hai Wang, Zhiliang Yan, Jingang Zhang and Yunfeng Nie
Remote Sens. 2026, 18(3), 432; https://doi.org/10.3390/rs18030432 - 29 Jan 2026
Viewed by 298
Abstract
Hyperspectral reconstruction (SR) refers to the computational process of generating high-dimensional hyperspectral images (HSI) from low-dimensional observations. However, the superior performance of most supervised learning-based reconstruction algorithms is predicated on the availability of fully labeled three-dimensional data. In practice, this requirement demands complex [...] Read more.
Hyperspectral reconstruction (SR) refers to the computational process of generating high-dimensional hyperspectral images (HSI) from low-dimensional observations. However, the superior performance of most supervised learning-based reconstruction algorithms is predicated on the availability of fully labeled three-dimensional data. In practice, this requirement demands complex optical paths with dual high-precision registrations and stringent calibration. To address this gap, we extend the fully supervised paradigm to a semi-supervised setting and propose SSHSR, a semi-supervised SR method for scenarios with limited spectral annotations. The core idea is to leverage spectrally aware mini-patches (SA-MP) as guidance and form region-level supervision from averaged spectra, so it can learn high-quality reconstruction without dense pixel-wise labels over the entire image. To improve reconstruction accuracy, we replace the conventional fixed-form Tikhonov physical layer with an optimizable version, which is then jointly trained with the deep network in an end-to-end manner. This enables the collaborative optimization of physical constraints and data-driven learning, thereby explicitly introducing learnable physical priors into the network. We also adopt a reconstruction network that combines spectral attention with spatial attention to strengthen spectral–spatial feature fusion and recover fine spectral details. Experimental results demonstrate that SSHSR outperforms existing state-of-the-art (SOTA) methods on several publicly available benchmark datasets, as well as on remote sensing and real-world scene data. On the GDFC remote sensing dataset, our method yields a 6.8% gain in PSNR and a 22.1% reduction in SAM. Furthermore, on our self-collected real-world scene dataset, our SSHSR achieves a 6.0% improvement in PSNR and a 11.9% decrease in SAM, confirming its effectiveness under practical conditions. Additionally, the model has only 1.59 M parameters, which makes it more lightweight than MST++ (1.62 M). This reduction in parameters lowers the deployment threshold while maintaining performance advantages, demonstrating its feasibility and practical value for real-world applications. Full article
Show Figures

Figure 1

29 pages, 2576 KB  
Review
A Semi-Supervised SVM-Firefly Hybrid for Rainfall Estimation from MSG Data
by Ouiza Boukendour, Mourad Lazri, Rafik Absi, Fethi Ouallouche, Karim Labadi, Youcef Attaf, Amar Belghit and Soltane Ameur
Atmosphere 2026, 17(2), 133; https://doi.org/10.3390/atmos17020133 - 26 Jan 2026
Viewed by 347
Abstract
In this paper, two improvements in precipitation classification have been performed. Supervised machine learning has demonstrated considerable performances in classification tasks. However, supervised machine learning can only be applied to labeled data. In some cases, large amounts of unlabeled data contain valuable information [...] Read more.
In this paper, two improvements in precipitation classification have been performed. Supervised machine learning has demonstrated considerable performances in classification tasks. However, supervised machine learning can only be applied to labeled data. In some cases, large amounts of unlabeled data contain valuable information for better classification. In the classification of precipitation intensities from satellite images, unlabeled data constitute the majority and remain largely unexplored. To exploit both labeled and unlabeled data, a Semi-Supervised Support Vector Machine (S3VM) is implemented as the first improvement for classification results. The labeling of the limited available data is derived from radar measurements covering a small portion of the Meteosat Second Generation Satellite observations. The results show that the S3VM model outperforms the standard SVM model, with up to a 15% improvement in classification accuracy compared to the standard SVM. To achieve the second improvement, the S3VM was combined with the Firefly Algorithm (FFA) to optimize its hyperparameters. This hybridization (S3VM-FFA) enabled an even more robust performance. A comparative study showed that the S3VM-FFA approach yielded highly satisfactory results, achieving a 17% improvement in classification compared to the SVM results. Based on these classifications, precipitation quantities at different scales are estimated. Similarly to the classification results, statistical evaluation parameters indicate that the S3VM-FFA outperforms both the standard SVM and the conventional S3VM. Full article
Show Figures

Figure 1

20 pages, 1567 KB  
Article
Deformable Pyramid Sparse Transformer for Semi-Supervised Driver Distraction Detection
by Qiang Zhao, Zhichao Yu, Jiahui Yu, Simon James Fong, Yuchu Lin, Rui Wang and Weiwei Lin
Sensors 2026, 26(3), 803; https://doi.org/10.3390/s26030803 - 25 Jan 2026
Viewed by 290
Abstract
Ensuring sustained driver attention is critical for intelligent transportation safety systems; however, the performance of data-driven driver distraction detection models is often limited by the high cost of large-scale manual annotation. To address this challenge, this paper proposes an adaptive semi-supervised driver distraction [...] Read more.
Ensuring sustained driver attention is critical for intelligent transportation safety systems; however, the performance of data-driven driver distraction detection models is often limited by the high cost of large-scale manual annotation. To address this challenge, this paper proposes an adaptive semi-supervised driver distraction detection framework based on teacher–student learning and deformable pyramid feature fusion. The framework leverages a limited amount of labeled data together with abundant unlabeled samples to achieve robust and scalable distraction detection. An adaptive pseudo-label optimization strategy is introduced, incorporating category-aware pseudo-label thresholding, delayed pseudo-label scheduling, and a confidence-weighted pseudo-label loss to dynamically balance pseudo-label quality and training stability. To enhance fine-grained perception of subtle driver behaviors, a Deformable Pyramid Sparse Transformer (DPST) module is integrated into a lightweight YOLOv11 detector, enabling precise multi-scale feature alignment and efficient cross-scale semantic fusion. Furthermore, a teacher-guided feature consistency distillation mechanism is employed to promote semantic alignment between teacher and student models at the feature level, mitigating the adverse effects of noisy pseudo-labels. Extensive experiments conducted on the Roboflow Distracted Driving Dataset demonstrate that the proposed method outperforms representative fully supervised baselines in terms of mAP@0.5 and mAP@0.5:0.95 while maintaining a balanced trade-off between precision and recall. These results indicate that the proposed framework provides an effective and practical solution for real-world driver monitoring systems under limited annotation conditions. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

17 pages, 3892 KB  
Article
Transformer-Driven Semi-Supervised Learning for Prostate Cancer Histopathology: A DINOv2–TransUNet Framework
by Rubina Akter Rabeya, Jeong-Wook Seo, Nam Hoon Cho, Hee-Cheol Kim and Heung-Kook Choi
Mach. Learn. Knowl. Extr. 2026, 8(2), 26; https://doi.org/10.3390/make8020026 - 23 Jan 2026
Viewed by 352
Abstract
Prostate cancer is diagnosed through a comprehensive study of histopathology slides, which takes time and requires professional interpretation. To minimize this load, we developed a semi-supervised learning technique that combines transformer-based representation learning and a custom TransUNet classifier. To capture a wide range [...] Read more.
Prostate cancer is diagnosed through a comprehensive study of histopathology slides, which takes time and requires professional interpretation. To minimize this load, we developed a semi-supervised learning technique that combines transformer-based representation learning and a custom TransUNet classifier. To capture a wide range of morphological structures without manual annotation, our method pretrains DINOv2 on 10,000 unlabeled prostate tissue patches. After receiving the transformer-derived features, a bespoke CNN-based decoder uses residual upsampling and carefully constructed skip connections to merge data from many spatial scales. Expert pathologists identified only 20% of the patches in the whole dataset; the remaining unlabeled samples were contributed by using a consistency-driven learning method that promoted reliable predictions across various augmentations. The model received precision and recall scores of 91.81% and 89.02%, respectively, and an accuracy of 93.78% on an additional test set. These results exceed the performance of a conventional U-Net and a baseline encoder–decoder network. All things considered, the localized CNN (Convolutional Neural Network) decoding and global transformer attention provide a reliable method for prostate cancer classification in situations with little annotated data. Full article
Show Figures

Graphical abstract

Back to TopTop