Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (341)

Search Parameters:
Keywords = histogram transformation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
35 pages, 5529 KB  
Article
Occasion-Based Clothing Classification Using Vision Transformer and Traditional Machine Learning Models
by Hanaa Alzahrani, Maram Almotairi and Arwa Basbrain
Computers 2026, 15(4), 249; https://doi.org/10.3390/computers15040249 - 17 Apr 2026
Viewed by 205
Abstract
Clothing classification by occasion is an important area in computer vision and artificial intelligence (AI). This task is particularly challenging because of the subtle visual similarities among clothing categories such as formal, party, and casual attire. Variations in color, fabric, patterns, and lighting [...] Read more.
Clothing classification by occasion is an important area in computer vision and artificial intelligence (AI). This task is particularly challenging because of the subtle visual similarities among clothing categories such as formal, party, and casual attire. Variations in color, fabric, patterns, and lighting further increase the complexity of this task. To address this challenge, we used the Fashionpedia dataset to create a balanced subset of 15,000 images. Specifically, we adopted two different methods for labeling these images: automated classification, which relies on category identifications (IDs) and components, and manual labeling performed by human annotators. We then implemented our preprocessing pipeline, which includes several steps: resizing, image normalization, background removal using segmentation masks, and class balancing. We benchmarked traditional models, including artificial neural networks (ANNs), support vector machines (SVMs), and k-nearest neighbors (KNNs), which use a histogram of oriented gradient (HOG) features, as well as deep learning models such as convolutional neural networks (CNNs), the Visual Geometry Group 16 (VGG16) model utilizing transfer learning, and the vision transformer (ViT) model, all evaluated using identical data splits and preprocessing procedures. The traditional models achieved moderate accuracy, ranging from 54% to 66%. In contrast, the ViT model achieved an accuracy of 81.78% with automated classification and 98.09% with manual labeling. This indicates that a higher label accuracy, along with the preprocessing steps used, significantly enhances the performance. Together, these factors improve the effectiveness of ViT in context-aware apparel classification and establish a reliable baseline for future research. Full article
(This article belongs to the Special Issue Machine Learning: Innovation, Implementation, and Impact)
Show Figures

Figure 1

18 pages, 5406 KB  
Article
ADC Histogram Features of Breast Cancer Brain Metastases as Candidate Imaging Biomarkers of Primary Tumor ER, PR, Ki-67, and Luminal Status
by Diba Saygılı Öz, Burcu Savran, Nazan Çiledağ, Özkan Ünal and Berna Karabulut
Diagnostics 2026, 16(8), 1154; https://doi.org/10.3390/diagnostics16081154 - 13 Apr 2026
Viewed by 336
Abstract
Background: Breast cancer brain metastases (BCBMs) are clinically challenging, and treatment decisions are influenced by tumor biology. Because receptor profiles may differ between primary breast tumors and brain metastases and brain biopsy may be impractical, non-invasive imaging biomarkers may provide useful biologic [...] Read more.
Background: Breast cancer brain metastases (BCBMs) are clinically challenging, and treatment decisions are influenced by tumor biology. Because receptor profiles may differ between primary breast tumors and brain metastases and brain biopsy may be impractical, non-invasive imaging biomarkers may provide useful biologic correlates. We evaluated whether diffusion-weighted imaging (DWI)-derived apparent diffusion coefficient (ADC) histogram metrics from BCBM were associated with primary tumor estrogen receptor (ER), progesterone receptor (PR), and human epidermal growth factor receptor 2 (HER2) status; the Ki-67 proliferation index; and luminal status. Methods: This retrospective exploratory single-center study included 72 adults with BCBM who underwent standardized 1.5T brain magnetic resonance imaging. The largest lesion in each patient was segmented on ADC maps in FireVoxel. ADC histogram features, including percentiles, were extracted. Using primary tumor biomarker status as the reference, candidate metrics were screened by univariable logistic regression. Parsimonious multivariable models included age, log-transformed lesion volume, and a single selected ADC percentile scaled by ×10. Discriminatory performance was assessed using area under the receiver operating characteristic curve (AUC); thresholds were derived with the Youden index. No external validation was performed. Results: Low-percentile ADC metrics were associated with ER positivity, PR positivity, and luminal disease, whereas no meaningful ADC histogram discrimination was observed for HER2. In multivariable models, ADC10×10 predicted ER positivity (odds ratio [OR] 0.441; AUC 0.847) and PR positivity (OR 0.478; AUC 0.819). Ki-67 positivity was best predicted by ADC75×10 (OR 3.095; AUC 0.905), although this finding should be interpreted cautiously. Luminal status (non-luminal vs. luminal) was predicted by ADC10×10 (OR 2.251; AUC 0.832). Conclusions: ADC histogram analysis from DWI in BCBM showed exploratory associations with primary tumor hormone receptor status and luminal subtype, but not HER2. These findings support ADC histogram features as candidate imaging biomarkers, but the Ki-67 result and all model performance estimates require cautious interpretation and independent external validation in multicenter cohorts. Full article
(This article belongs to the Section Medical Imaging and Theranostics)
Show Figures

Graphical abstract

20 pages, 3303 KB  
Article
Revisiting Remote Sensing Image Dehazing via a Dynamic Histogram-Sorted Transformer
by Naiwei Chen, Xin He, Shengyuan Li, Fengning Liu, Haoyi Lv, Haowei Peng and Yuebu Qubie
Remote Sens. 2026, 18(7), 1040; https://doi.org/10.3390/rs18071040 - 30 Mar 2026
Viewed by 321
Abstract
Remote sensing images are highly susceptible to spatially non-uniform haze under complex atmospheric conditions, leading to contrast degradation and structural detail loss. Moreover, remote sensing scenes usually exhibit complex spatial structures, highly uneven haze distribution, and significant statistical variability, which further increases the [...] Read more.
Remote sensing images are highly susceptible to spatially non-uniform haze under complex atmospheric conditions, leading to contrast degradation and structural detail loss. Moreover, remote sensing scenes usually exhibit complex spatial structures, highly uneven haze distribution, and significant statistical variability, which further increases the difficulty of haze removal. To address this issue, we revisit the haze degradation mechanism of remote sensing imagery and propose a dynamic histogram-sorted Transformer dehazing method from the perspectives of statistical distribution modeling and region-adaptive restoration. Specifically, a Histogram-Sorted Adaptive Attention is designed to map spatial features into the statistical distribution domain through a dynamic histogram sorting mechanism, enabling explicit discrimination and precise modeling of regions with different haze densities. Meanwhile, a Perception-Adaptive Feed-Forward Network is constructed, which incorporates a stable routing-based mixture-of-experts mechanism to adaptively select restoration strategies according to local texture characteristics and global haze density, thereby significantly enhancing the adaptability of the model in complex remote sensing scenarios. Extensive experimental results demonstrate that the proposed method achieves superior performance over existing approaches across multiple remote sensing benchmark datasets, effectively improving both visual quality and robustness of remote sensing imagery. Full article
Show Figures

Figure 1

18 pages, 12071 KB  
Article
A Novel Reversible Image Camouflaging Method Based on Lossless Matrix Transformation
by Gizem Dursun Demir and Ufuk Özkaya
Mathematics 2026, 14(7), 1111; https://doi.org/10.3390/math14071111 - 26 Mar 2026
Viewed by 318
Abstract
Image encryption methods aim to transform a secret image into a noise-like, texture-like image. Since this behavior of the encrypted image indicates that it is encrypted, it provokes a large number of attacks. One of the most effective methods to counter this threat [...] Read more.
Image encryption methods aim to transform a secret image into a noise-like, texture-like image. Since this behavior of the encrypted image indicates that it is encrypted, it provokes a large number of attacks. One of the most effective methods to counter this threat is to protect the information by transforming the original image into a new, meaningful image. The bottleneck of this approach is that the new image in which the information is embedded must have a high visual quality that is indistinguishable from the real image. Another critical requirement is obtaining the original image without loss. In this paper, we propose a reversible image camouflage method based on lossless matrix transformation and two-dimensional wavelet transformation. Random matrix perturbation is introduced and applied as an effective method for the lossless transformation of low-frequency or flat regions. The proposed method was applied to different datasets for performance analysis. The PSNR values of the plain/camouflage image pair are above 55 dB, and the SSIM values obtained by our method are very close to 0.9999 on these datasets. The experimental results demonstrate that the method’s performance is independent of the content of the plain/target image and of the fragment size. Furthermore, in cases where the target image is specifically chosen, PSNR values exceed 58 dB. Additionally, the efficacy of the method in generating camouflage images has been demonstrated through histogram analysis and performance analysis in the low- and high-frequency regions. Full article
(This article belongs to the Section E1: Mathematics and Computer Science)
Show Figures

Figure 1

20 pages, 3154 KB  
Article
A Data-Centric Algorithmic Pipeline for Enhancing Cardiac MRI Segmentation Using ViTUNeT and Quality-Aware Filtering
by Salvador de Haro, Jesús Cámara, Pilar González-Férez, José Manuel García and Gregorio Bernabé
Algorithms 2026, 19(3), 200; https://doi.org/10.3390/a19030200 - 6 Mar 2026
Viewed by 327
Abstract
The performance of deep-learning-based segmentation models is strongly dependent on the quality of the input data, which is frequently heterogeneous or degraded in real-world medical imaging scenarios. This work presents a data-centric algorithmic pipeline designed to improve cardiac MRI segmentation accuracy through systematic [...] Read more.
The performance of deep-learning-based segmentation models is strongly dependent on the quality of the input data, which is frequently heterogeneous or degraded in real-world medical imaging scenarios. This work presents a data-centric algorithmic pipeline designed to improve cardiac MRI segmentation accuracy through systematic image enhancement and automatic slice-quality filtering. The proposed method is formalized as deterministic algorithm that combines image processing and supervised learning components. The approach integrates a contrast- and structure-preserving enhancement stage, based on bilateral filtering and adaptive histogram equalization, with a quality-aware selection algorithm. Slice quality is assessed using anatomical attributes extracted via YOLOv11s-based localization and a supervised classification model trained to identify diagnostically reliable images. When applied to transformer-based segmentation architectures such as ViTUNeT, the pipeline yields consistent improvements across all evaluation metrics without increasing model complexity or training cost. These findings emphasize the importance of algorithmic data curation as an effective strategy for enhancing robustness and stability in deep-learning segmentation pipelines and demonstrate the broader applicability of the proposed approach to computer-vision tasks involving heterogeneous or low-quality image datasets. Full article
(This article belongs to the Special Issue AI-Powered Biomedical Image Analysis)
Show Figures

Figure 1

18 pages, 5195 KB  
Article
Computational Ghost Imaging Encryption for Multiple Images Based on Compressed Sensing and Block Scrambling
by Zhipeng Wang, Jiahuan Yang, Ruizhi Ge, Yingying Zhang and Yi Qin
Information 2026, 17(3), 239; https://doi.org/10.3390/info17030239 - 1 Mar 2026
Viewed by 358
Abstract
To achieve high capacity, high speed, and secure image transmission, we propose a multi-image computational ghost imaging (CGI)-based encryption scheme that integrates compressed sensing (CS), block scrambling, and dynamic-salt-driven bidirectional XOR diffusion. First, multiple images are partitioned into 8 × 8 pixel blocks, [...] Read more.
To achieve high capacity, high speed, and secure image transmission, we propose a multi-image computational ghost imaging (CGI)-based encryption scheme that integrates compressed sensing (CS), block scrambling, and dynamic-salt-driven bidirectional XOR diffusion. First, multiple images are partitioned into 8 × 8 pixel blocks, and their spatial structure is disrupted through random scrambling. The scrambled composite image then undergoes pixel-level encryption via two-round bidirectional XOR diffusion, using session-unique keys derived from SHA-256-based dynamic salt, eliminating the statistical characteristics of the original images. Subsequently, each pixel block is subjected to both Gaussian CS and Hadamard-based CGI measurements in parallel, achieving dual-mode compressive encryption and enhancing robustness through measurement redundancy. Finally, only the scrambling key, the XOR-diffusion key, and the compressed measurements are stored; the original image information is thus transformed into unrecognizable measurement data. During the decryption process, the Fast Iterative Shrinkage-Thresholding Algorithm (FISTA) with a Discrete Cosine Transform (DCT) sparse basis is employed for dual-sparse reconstruction from the compressed measurements, recovering the encrypted composite image. An inverse XOR operation is then applied to remove the pixel-level diffusion, followed by block reordering using the scrambling key to restore the original images. Experimental results demonstrate that the proposed scheme enables efficient and secure multi-image transmission while maintaining high decrypted image quality. Security analysis indicates that the scheme possesses high key sensitivity, effectively resisting chosen-plaintext attacks. Histogram uniformity analysis and cropping attack resistance experiments further confirm its excellent statistical security and robustness. Full article
(This article belongs to the Section Information Processes)
Show Figures

Figure 1

27 pages, 4803 KB  
Article
Enhancing Short-Term Wind Energy Forecasting with XGBoost and Conformal Prediction for Robust Uncertainty Quantification
by Rabelani Innocent Nthangeni, Caston Sigauke, Thakhani Ravele and Thinawanga Hangwani Tshisikhawe
Computation 2026, 14(3), 56; https://doi.org/10.3390/computation14030056 - 1 Mar 2026
Viewed by 422
Abstract
This paper presents probabilistic wind energy forecasting using quantile regression averaging combined with a conformal prediction modelling framework. The study uses data from Eskom, South Africa’s power utility company. The data is from April 2019 to November 2023. A partial linear additive quantile [...] Read more.
This paper presents probabilistic wind energy forecasting using quantile regression averaging combined with a conformal prediction modelling framework. The study uses data from Eskom, South Africa’s power utility company. The data is from April 2019 to November 2023. A partial linear additive quantile regression (PLAQR) averaging method is used to combine forecasts from two competing forecasting models: eXtreme Gradient Boosting (XGBoost) and Principal Component Regression (PCR). To compare the predictive abilities of the models, two data splits are used: 80%, 10% and 10% for the first set, and 85%, 10% and 5% for the second set, for training, validation and testing, respectively. Empirical results suggest that the combined predictions from PLAQR perform better than the individual models, significantly improving calibration and accuracy. The proposed combination has the smallest root mean square error (RMSE) and the highest probability of change in direction (POCID). The combination captures nonlinearities and produces well-calibrated probabilistic results. Probability integral transform histograms validate this. This performance gain reflected the importance of data volume. This is reinforced by the fact that the PLAQR model, which combines the benefits of tree-based approaches and linear models, is a robust modelling approach for reliable renewable energy forecasting. Future research directions should consider more varied ensembles. Full article
(This article belongs to the Section Computational Engineering)
Show Figures

Figure 1

19 pages, 1049 KB  
Review
Image-Guided Adaptive Brachytherapy for Uterine Cancer: A Comprehensive Review
by Yi-Ching Chen and Chi-Yuan Yeh
Cancers 2026, 18(4), 693; https://doi.org/10.3390/cancers18040693 - 20 Feb 2026
Viewed by 622
Abstract
Background/Objectives: Image-guided adaptive brachytherapy (IGABT) has transformed the standard of care for locally advanced cervical cancer (LACC), enabling volumetric target definition and dose–volume histogram (DVH)-based planning to improve pelvic tumor control while limiting severe late toxicity. Methods: A comprehensive literature search [...] Read more.
Background/Objectives: Image-guided adaptive brachytherapy (IGABT) has transformed the standard of care for locally advanced cervical cancer (LACC), enabling volumetric target definition and dose–volume histogram (DVH)-based planning to improve pelvic tumor control while limiting severe late toxicity. Methods: A comprehensive literature search of PubMed/MEDLINE and Embase was done for articles published up to August 2024, using combinations of the following keywords and Medical Subject Heading (MeSH) terms: “cervical cancer”, “endometrial cancer”, “vaginal cancer”, “uterine neoplasms”, “brachytherapy”, “high-dose-rate”, “image-guided”, “MRI-guided”, “3D brachytherapy”, “IGABT”, “interstitial”, “locoregional control”, “toxicity”, “quality of life”, and “patient-reported outcomes”. Results: We summarized the contemporary evidence on IGABT for cervical, endometrial, and primary or recurrent vaginal cancers, focusing on local control, survival, late morbidity, and patient-reported outcomes. We described the key target volume concepts (gross tumor volume, high- and intermediate-risk clinical target volumes), and the role of MRI-, CT-, and ultrasound-based planning with intracavitary, intracavitary–interstitial, and interstitial applicators. Conclusions: Image-guided adaptive brachytherapy has redefined the standard of care for the management of locally advanced cervical cancer. Through the integration of volumetric target concepts, DVH-based dose reporting, and advanced imaging, IGABT has enabled consistent dose escalation to the residual tumor while accounting for organ-at-risk constraints, resulting in high local control rates and reduced severe morbidity compared with historical 2D brachytherapy. Full article
Show Figures

Figure 1

22 pages, 1730 KB  
Article
Toward a Hybrid Intrusion Detection Framework for IIoT Using a Large Language Model
by Musaad Algarni, Mohamed Y. Dahab, Abdulaziz A. Alsulami, Badraddin Alturki and Raed Alsini
Sensors 2026, 26(4), 1231; https://doi.org/10.3390/s26041231 - 13 Feb 2026
Viewed by 619
Abstract
The widespread connectivity of the Industrial Internet of Things (IIoT) improves the efficiency and functionality of connected devices. However, it also raises serious concerns about cybersecurity threats. Implementing an effective intrusion detection system (IDS) for IIoT is challenging due to heterogeneous data, high [...] Read more.
The widespread connectivity of the Industrial Internet of Things (IIoT) improves the efficiency and functionality of connected devices. However, it also raises serious concerns about cybersecurity threats. Implementing an effective intrusion detection system (IDS) for IIoT is challenging due to heterogeneous data, high feature dimensionality, class imbalance, and the risk of data leakage during evaluation. This paper presents a leakage-safe hybrid intrusion detection framework that combines text-based and numerical network flow features in an IIoT environment. Each network flow is converted into a short text description and encoded using a frozen Large Language Model (LLM) called the Bidirectional Encoder Representations from Transformers (BERT) model to obtain fixed semantic embeddings, while numerical traffic features are standardized in parallel. To improve class separation, class prototypes are computed in Principal Component Analysis (PCA) space, and cosine similarity scores for these prototypes are added to the feature set. Class imbalance is handled only in the training data using the Synthetic Minority Over-sampling Technique (SMOTE). A Random Forest (RF) is used to select the top features, followed by a Histogram-based Gradient Boosting (HGB) classifier for final prediction. The proposed framework is evaluated on the Edge-IIoTset and ToN_IoT datasets and achieves promising results. Empirically, the framework attains 98.19% accuracy on Edge-IIoTset and 99.15% accuracy on ToN_IoT, indicating robust, leakage-safe performance. Full article
Show Figures

Figure 1

19 pages, 2617 KB  
Article
Topic-Modeling Guided Semantic Clustering for Enhancing CNN-Based Image Classification Using Scale-Invariant Feature Transform and Block Gabor Filtering
by Natthaphong Suthamno and Jessada Tanthanuch
J. Imaging 2026, 12(2), 70; https://doi.org/10.3390/jimaging12020070 - 9 Feb 2026
Viewed by 413
Abstract
This study proposes a topic-modeling guided framework that enhances image classification by introducing semantic clustering prior to CNN training. Images are processed through two key-point extraction pipelines: Scale-Invariant Feature Transform (SIFT) with Sobel edge detection and Block Gabor Filtering (BGF), to obtain local [...] Read more.
This study proposes a topic-modeling guided framework that enhances image classification by introducing semantic clustering prior to CNN training. Images are processed through two key-point extraction pipelines: Scale-Invariant Feature Transform (SIFT) with Sobel edge detection and Block Gabor Filtering (BGF), to obtain local feature descriptors. These descriptors are clustered using K-means to build a visual vocabulary. Bag of Words histograms then represent each image as a visual document. Latent Dirichlet Allocation is applied to uncover latent semantic topics, generating coherent image clusters. Cluster-specific CNN models, including AlexNet, GoogLeNet, and several ResNet variants, are trained under identical conditions to identify the most suitable architecture for each cluster. Two topic guided integration strategies, the Maximum Proportion Topic (MPT) and the Weight Proportion Topic (WPT), are then used to assign test images to the corresponding specialized model. Experimental results show that both the SIFT-based and BGF-based pipelines outperform non-clustered CNN models and a baseline method using Incremental PCA, K-means, Same-Cluster Prediction, and unweighted Ensemble Voting. The SIFT pipeline achieves the highest accuracy of 95.24% with the MPT strategy, while the BGF pipeline achieves 93.76% with the WPT strategy. These findings confirm that semantic structure introduced through topic modeling substantially improves CNN classification performance. Full article
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)
Show Figures

Figure 1

25 pages, 51444 KB  
Article
Local Contrast Enhancement in Digital Images Using a Tunable Modified Hyperbolic Tangent Transformation
by Camilo E. Echeverry and Manuel G. Forero
Mathematics 2026, 14(3), 571; https://doi.org/10.3390/math14030571 - 5 Feb 2026
Viewed by 455
Abstract
Low contrast is a frequent challenge in image analysis, especially within medical imaging and highly saturated scenes. To address this issue, we present a nonlinear transformation for local contrast enhancement in digital images. Our method adapts the hyperbolic tangent function using two parameters: [...] Read more.
Low contrast is a frequent challenge in image analysis, especially within medical imaging and highly saturated scenes. To address this issue, we present a nonlinear transformation for local contrast enhancement in digital images. Our method adapts the hyperbolic tangent function using two parameters: one to select the intensity range for modification and another to control the degree of enhancement. This approach outperforms conventional histogram-based techniques such as histogram equalization and specification in local contrast enhancement, without increasing computational cost, and produces smooth, artifact-free results in user-defined regions of interest. In addition, the proposed method was compared with CLAHE in MRIs, showing that, unlike CLAHE, the proposed method does not enhance the noise present in the background of the image. Furthermore, in deep learning contexts where dataset size is often limited, our method could serve as an effective data augmentation tool—generating varied contrast images while preserving anatomical structures, which improves neural network training for brain tumor detection in magnetic resonance imaging. The ability to manipulate local contrast may offer a pathway toward better interpretability of convolutional neural networks, as targeted contrast adjustments allow researchers to probe model sensitivity and enhance the explainability of classification and detection mechanisms. Full article
(This article belongs to the Special Issue Data Mining and Algorithms Applied in Image Processing)
Show Figures

Figure 1

52 pages, 9165 KB  
Article
A Hybrid Deep Learning Framework for Automated Dental Disorder Diagnosis from X-Ray Images
by A. A. Abd El-Aziz, Mohammed Elmogy, Mahmood A. Mahmood and Sameh Abd El-Ghany
J. Clin. Med. 2026, 15(3), 1076; https://doi.org/10.3390/jcm15031076 - 29 Jan 2026
Viewed by 478
Abstract
Background: Dental disorders, such as cavities, periodontal disease, and periapical infections, remain major global health issues, often resulting in pain, tooth loss, and systemic complications if not identified early. Traditional diagnostic methods rely heavily on visual inspection and manual interpretation of panoramic X-ray [...] Read more.
Background: Dental disorders, such as cavities, periodontal disease, and periapical infections, remain major global health issues, often resulting in pain, tooth loss, and systemic complications if not identified early. Traditional diagnostic methods rely heavily on visual inspection and manual interpretation of panoramic X-ray images by dental professionals, making them time-consuming, subjective, and less accessible in resource-limited settings. Objectives: Accurate and timely diagnosis is vital for effective treatment and prevention of disease progression, reducing healthcare costs and patient discomfort. Recent advances in deep learning (DL) have demonstrated remarkable potential to automate and improve the precision of dental diagnostics by objectively analyzing panoramic, periapical, and bitewing X-rays. Methods: In this research, a hybrid feature-fusion framework is proposed. It integrates handcrafted Histogram of Oriented Gradients (HOG) features with deep representations from DenseNet-201 and the Shifted Window (Swin) Transformer models. Sequential dependencies among the fused features were learned utilizing the Long Short-Term Memory (LSTM) classifier. The framework was evaluated on the Dental Radiography Analysis and Diagnosis (DRAD) dataset following preprocessing steps, including resizing, normalization, Contrast Limited Adaptive Histogram Equalization (CLAHE) enhancement, and image cropping. Results: The proposed LSTM-based hybrid model achieved 96.47% accuracy, 91.76% specificity, 94.92% precision, 91.76% recall, and 93.14% F1-score. Conclusions: The proposed framework offers flexibility, interpretability, and strong empirical performance, making it suitable for various image-based recognition applications and serving as a reproducible framework for future research on hybrid feature fusion and sequence-based classification. Full article
(This article belongs to the Special Issue Clinical Advances in Cancer Imaging)
Show Figures

Figure 1

33 pages, 2852 KB  
Article
Robust Activity Recognition via Redundancy-Aware CNNs and Novel Pooling for Noisy Mobile Sensor Data
by Bnar Azad Hamad Ameen and Sadegh Abdollah Aminifar
Sensors 2026, 26(2), 710; https://doi.org/10.3390/s26020710 - 21 Jan 2026
Viewed by 481
Abstract
This paper proposes a robust convolutional neural network (CNN) architecture for human activity recognition (HAR) using smartphone accelerometer data, evaluated on the WISDM dataset. We introduce two novel pooling mechanisms—Pooling A (Extrema Contrast Pooling (ECP)) and Pooling B (Center Minus Variation (CMV))—that enhance [...] Read more.
This paper proposes a robust convolutional neural network (CNN) architecture for human activity recognition (HAR) using smartphone accelerometer data, evaluated on the WISDM dataset. We introduce two novel pooling mechanisms—Pooling A (Extrema Contrast Pooling (ECP)) and Pooling B (Center Minus Variation (CMV))—that enhance feature discrimination and noise robustness. ECP emphasizes sharp signal transitions through a nonlinear penalty based on the squared range between extrema, while CMV Pooling penalizes local variability by subtracting the standard deviation, improving resilience to noise. Input data are normalized to the [0, 1] range to ensure bounded and interpretable pooled outputs. The proposed framework is evaluated in two separate configurations: (1) a 1D CNN applied to raw tri-axial sensor streams with the proposed pooling layers, and (2) a histogram-based image encoding pipeline that transforms segment-level sensor redundancy into RGB representations for a 2D CNN with fully connected layers. Ablation studies show that histogram encoding provides the largest improvement, while the combination of ECP and CMV further enhances classification performance. Across six activity classes, the 2D CNN system achieves up to 96.84% weighted classification accuracy, outperforming baseline models and traditional average pooling. Under Gaussian, salt-and-pepper, and mixed noise conditions, the proposed pooling layers consistently reduce performance degradation, demonstrating improved stability in real-world sensing environments. These results highlight the benefits of redundancy-aware pooling and histogram-based representations for accurate and robust mobile HAR systems. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

23 pages, 13473 KB  
Article
Automatic Threshold Selection Guided by Maximizing Homologous Isomeric Similarity Under Unified Transformation Toward Unimodal Distribution
by Yaobin Zou, Wenli Yu and Qingqing Huang
Electronics 2026, 15(2), 451; https://doi.org/10.3390/electronics15020451 - 20 Jan 2026
Viewed by 1546
Abstract
Traditional thresholding methods are often tailored to specific histogram patterns, making it difficult to achieve robust segmentation across diverse images exhibiting non-modal, unimodal, bimodal, or multimodal distributions. To address this limitation, this paper proposes an automatic thresholding method guided by maximizing homologous isomeric [...] Read more.
Traditional thresholding methods are often tailored to specific histogram patterns, making it difficult to achieve robust segmentation across diverse images exhibiting non-modal, unimodal, bimodal, or multimodal distributions. To address this limitation, this paper proposes an automatic thresholding method guided by maximizing homologous isomeric similarity under a unified transformation toward unimodal distribution. The primary objective is to establish a generalized selection criterion that functions independently of the input histogram’s pattern. The methodology employs bilateral filtering, non-maximum suppression, and Sobel operators to transform diverse histogram patterns into a unified, right-skewed unimodal distribution. Subsequently, the optimal threshold is determined by maximizing the normalized Renyi mutual information between the transformed edge image and binary contour images extracted at varying levels. Experimental validation on both synthetic and real-world images demonstrates that the proposed method offers greater adaptability and higher accuracy compared to representative thresholding and non-thresholding techniques. The results show a significant reduction in misclassification errors and improved correlation metrics, confirming the method’s effectiveness as a unified thresholding solution for images with non-modal, unimodal, bimodal, or multimodal histogram patterns. Full article
(This article belongs to the Special Issue Image Processing and Pattern Recognition)
Show Figures

Figure 1

19 pages, 2960 KB  
Article
Gabor Transform-Based Deep Learning System Using CNN for Melanoma Detection
by S. Deivasigamani, C. Senthilpari, Siva Sundhara Raja. D, A. Thankaraj, G. Narmadha and K. Gowrishankar
Computers 2026, 15(1), 54; https://doi.org/10.3390/computers15010054 - 13 Jan 2026
Viewed by 540
Abstract
Melanoma is highly dangerous and can spread rapidly to other parts of the body. It has an increasing fatality rate among different types of cancer. Timely detection of skin malignancies can reduce overall mortality. Therefore, clinical screening methods require more time and accuracy [...] Read more.
Melanoma is highly dangerous and can spread rapidly to other parts of the body. It has an increasing fatality rate among different types of cancer. Timely detection of skin malignancies can reduce overall mortality. Therefore, clinical screening methods require more time and accuracy for diagnosis. An automated, computer-aided system would facilitate earlier melanoma detection, thereby increasing patient survival rates. This paper identifies melanoma images using a Convolutional Neural Network. Skin images are preprocessed using Histogram Equalization and Gabor transforms. A Gabor filter-based Convolutional Neural Network (CNN) classifier trains and classifies the extracted features. We adopt Gabor filters because they are bandpass filters that transform a pixel into a multi-resolution kernel matrix, providing detailed information about the image. This study suggests a method with accuracy, sensitivity, and specificity of 98.58%, 98.66%, and 98.75%, respectively. This research supports SDGs 3 and 4 by facilitating early melanoma detection and enhancing AI-driven medical education. Full article
Show Figures

Figure 1

Back to TopTop