Previous Article in Journal
Assessing Change in Stone Burden on Baseline and Follow-Up CT: Radiologist and Radiomics Evaluations
Previous Article in Special Issue
ODDM: Integration of SMOTE Tomek with Deep Learning on Imbalanced Color Fundus Images for Classification of Several Ocular Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification

by
Jatsada Singthongchai
and
Tanachapong Wangkhamhan
*
Department of Computer Science and Information Technology, Faculty of Science and Health Technology, Kalasin University, Kalasin 46000, Thailand
*
Author to whom correspondence should be addressed.
J. Imaging 2026, 12(1), 14; https://doi.org/10.3390/jimaging12010014
Submission received: 18 November 2025 / Revised: 18 December 2025 / Accepted: 26 December 2025 / Published: 28 December 2025
(This article belongs to the Special Issue Advances in Machine Learning for Medical Imaging Applications)

Abstract

This study presents a controlled benchmarking analysis of min–max scaling, Z-score normalization, and an adaptive preprocessing pipeline that combines percentile-based ROI cropping with histogram standardization. The evaluation was conducted across four public chest X-ray (CXR) datasets and three convolutional neural network architectures under controlled experimental settings. The adaptive pipeline generally improved accuracy, F1-score, and training stability on datasets with relatively stable contrast characteristics while yielding limited gains on MIMIC-CXR due to strong acquisition heterogeneity. Ablation experiments showed that histogram standardization provided the primary performance contribution, with ROI cropping offering complementary benefits, and the full pipeline achieving the best overall performance. The computational overhead of the adaptive preprocessing was minimal (+6.3% training-time cost; 5.2 ms per batch). Friedman–Nemenyi and Wilcoxon signed-rank tests confirmed that the observed improvements were statistically significant across most dataset–model configurations. Overall, adaptive normalization is positioned not as a novel algorithmic contribution, but as a practical preprocessing design choice that can enhance cross-dataset robustness and reliability in chest X-ray classification workflows.

1. Introduction

Chest radiography (CXR) is a widely used imaging modality due to its rapid acquisition, low cost, and non-invasive nature, supporting diagnosis across a broad range of thoracic diseases [1,2]. With advances in artificial intelligence (AI), deep learning-based CXR analysis has shown promise in improving triage efficiency and diagnostic support, particularly in high-volume and resource-limited clinical settings [1,3,4].
Despite these advances, the generalization of deep learning (DL) models across institutions remains limited. Two closely related factors contribute to this challenge. First, resizing high-resolution CXRs to standard network input sizes (e.g., 256 × 256 pixels) reduces the relative lung-to-image ratio, allowing non-diagnostic background structures to dominate learned representations and obscure subtle abnormalities [5,6]. Second, substantial domain shift arises from heterogeneity in imaging devices, acquisition protocols, contrast, brightness, and preprocessing pipelines across institutions, which can significantly degrade model performance when deployed beyond the training domain [7,8].
Image normalization is a central yet under-benchmarked factor in addressing these issues. While conventional strategies such as min–max scaling and Z-score normalization are widely adopted, their effectiveness under cross-dataset variability has rarely been evaluated in a controlled and systematic manner. Prior studies have explored adaptive instance normalization [9], contrastive domain alignment [10,11,12], multimodal integration with clinical metadata [2], and histogram-based or localized intensity normalization techniques [13,14,15,16]. However, these approaches are typically evaluated within single datasets or fixed model architectures, making it difficult to isolate the role of normalization from confounding factors such as dataset scale, model capacity, or training protocol.
To address this gap, this study conducted a systematic benchmarking analysis of three normalization strategies—min–max scaling, Z-score normalization, and an adaptive preprocessing pipeline combining percentile-based ROI cropping with histogram standardization. The evaluation was performed across four public CXR datasets, including CheXpert [17], MIMIC-CXR [18], and ChestX-ray14 [6], using three convolutional neural network architectures: a lightweight CNN, EfficientNet-B0, and MobileNetV2, under controlled sampling and identical training conditions.
This work is positioned as a controlled benchmarking study rather than a methodological innovation. The main contributions are threefold:
(1)
It establishes a controlled cross-dataset and cross-architecture evaluation framework for comparing normalization strategies;
(2)
It quantifies the impact of normalization choices on cross-domain generalization, training stability, and performance consistency, with particular emphasis on lightweight architectures such as MobileNetV2; and
(3)
It provides a statistically grounded comparison using Friedman–Nemenyi and Wilcoxon signed-rank tests to clarify when adaptive normalization yields meaningful performance gains over conventional approaches.
The remainder of this paper is organized as follows. Section 2 reviews related work on normalization and domain generalization in CXR analysis. Section 3 describes the datasets, normalization techniques, and experimental methodology. Section 4 presents the experimental results. Section 5 discusses the implications and limitations of the findings, and Section 6 concludes the study and outlines directions for future work.

2. Background and Related Work

Robust evaluation of normalization strategies for chest X-ray (CXR) classification requires datasets that reflect meaningful diversity in patient populations, acquisition protocols, and imaging conditions. Prior studies have demonstrated that deep learning models for CXR are highly sensitive to domain shifts introduced by heterogeneity in imaging sources and preprocessing pipelines [7,8,19]. Accordingly, the datasets and preprocessing techniques reviewed in this section provide the foundation for cross-dataset benchmarking of normalization methods.

2.1. Datasets

2.1.1. ChestX-ray14

ChestX-ray14 contains over 112,000 frontal CXR images from more than 30,000 patients, labeled with fourteen thoracic disease categories [6,20]. Due to label noise arising from automated report mining, it is commonly used to evaluate model robustness under imperfect supervision [8,21].

2.1.2. CheXpert

CheXpert includes more than 220,000 images from approximately 65,000 patients and explicitly models diagnostic uncertainty in its labels [17]. This design supports the evaluation of calibration, robustness, and uncertainty-aware learning in CXR classification [15,22].

2.1.3. MIMIC-CXR

MIMIC-CXR comprises over 370,000 CXR images collected across multiple departments and imaging devices, introducing substantial acquisition heterogeneity [18]. This diversity makes it a key benchmark for assessing cross-domain generalization in realistic clinical settings [23].

2.1.4. Pediatric Chest X-Ray (Kermany Dataset)

The pediatric CXR dataset contains 5863 images labeled as normal, bacterial pneumonia, or viral pneumonia [24]. Despite its smaller scale, it represents a distinct anatomical domain and is frequently used to study transfer learning and adult-to-pediatric generalization [25,26].

2.2. Preprocessing Techniques

Preprocessing is essential for mitigating domain shift caused by variations in anatomy, acquisition parameters, and institutional imaging protocols [7,11,27]. Among preprocessing strategies, normalization plays a central role in stabilizing intensity distributions, while data augmentation provides complementary benefits through geometric and intensity transformations [28].

2.2.1. Normalization

Normalization aims to reduce intensity variability induced by scanner differences, exposure settings, and patient conditions. Conventional techniques such as min–max scaling and Z-score normalization remain widely used due to their simplicity and compatibility with convolutional architectures [29]. However, their effectiveness is limited in multi-source settings with heterogeneous contrast distributions.
Recent work has explored adaptive and spatially aware normalization schemes, including adaptive instance normalization and localized histogram-based methods, which demonstrate improved robustness in cross-domain and self-supervised learning pipelines [14,16,30]. These studies highlight the importance of region-focused intensity transformations for heterogeneous CXR datasets.

2.2.2. Min–Max Scaling as a Baseline

Min–max scaling linearly maps pixel intensities to a fixed range and is commonly used as a baseline due to its low computational cost [31,32]. Nevertheless, it performs poorly under cross-dataset variability because it does not account for differences in underlying intensity distributions [33,34].

2.2.3. Z-Score Normalization as a Standard Baseline

Z-score normalization standardizes intensities to zero mean and unit variance and has demonstrated greater resilience to device heterogeneity than min–max scaling in multi-center studies [35]. It is also associated with improved calibration and reduced performance variance in domain-adaptive frameworks [13,36].

2.3. Model Architectures

To assess normalization effects under realistic deployment constraints, three convolutional architectures with varying computational complexity are commonly adopted: lightweight CNNs for resource-constrained environments, EfficientNet-B0 for balanced accuracy and efficiency, and MobileNetV2 for mobile and embedded applications [37,38]. Training these architectures under identical optimization settings enables the isolation of normalization effects from architectural factors.

2.4. Region of Interest and Signal-to-Noise Ratio

Resizing high-resolution CXRs to standard input resolutions can reduce the lung-to-image ratio and obscure subtle pulmonary findings [1,6]. ROI-based cropping techniques mitigate this effect by preserving diagnostically relevant regions. Percentile-based cropping defined in relative coordinate space has been shown to provide stable anatomical coverage across datasets with varying resolutions and aspect ratios, supporting improved cross-dataset generalization [11].

2.5. Domain Adaptation and Histogram Standardization

Domain shifts caused by inter-dataset variability in brightness and contrast can significantly degrade model performance [11,18]. Histogram standardization aligns image intensity statistics to a reference distribution and has been shown to reduce inter-dataset variability, particularly when combined with region-focused preprocessing strategies [16,39].

2.6. Comparative Analysis of Related Work

Recent studies emphasize that normalization choices substantially influence cross-domain robustness in CXR classification [35,36]. While advances in segmentation-driven pipelines, federated learning, and model compression address complementary challenges [15,40,41], systematic comparisons of normalization strategies across datasets and architectures remain limited. This gap motivates the benchmarking framework adopted in this study.
Building on this body of work, our study presents a systematic benchmarking of normalization strategies across multiple datasets and CNN architectures, as summarized in Table 1.

2.7. Transformer and Foundation Model Approaches

Transformer-based and foundation model approaches, including vision–language pretraining and hybrid CNN–transformer architectures, have demonstrated strong transfer and zero-shot performance in CXR analysis [22,48,49,50,51]. Although this study focused on convolutional backbones for controlled benchmarking, the preprocessing issues examined—ROI selection, intensity normalization, and cross-domain robustness, remain directly relevant to transformer-based and foundation model pipelines, which also depend on stable input distributions for reliable downstream performance.

3. Methodology

3.1. Dataset Description

This study utilized four publicly available chest X-ray datasets that represent diverse imaging conditions, patient populations, and diagnostic labels. Three large-scale adult datasets, ChestX-ray14, CheXpert, and MIMIC-CXR, were uniformly sampled to 16,000 images each in order to formulate a controlled evaluation environment where normalization effects can be compared under identical sample sizes. The pediatric Chest-X-ray-Pneumonia dataset contains 5863 images, and the entire set was included due to its smaller size. For the pediatric dataset, class-proportion balancing was applied to mitigate label imbalance before training.
Standardizing the dataset sizes helps ensure that performance differences across normalization methods are not influenced by variations in dataset scale. This controlled sampling strategy follows recommendations from prior evaluation frameworks that emphasize reproducibility and statistical fairness when comparing preprocessing approaches in medical imaging [7,35].
A summary of the datasets used in this study is provided in Table 2.

3.2. Image Preprocessing Techniques

Normalization is a critical component in stabilizing the training dynamics and improving the generalization of deep learning models for chest X-ray (CXR) analysis. Variations in acquisition devices, patient anatomy, and exposure settings introduce substantial heterogeneity across datasets, making preprocessing essential for cross-domain robustness [52]. In this study, three normalization techniques were evaluated systematically: scaling normalization, Z-score normalization, and the proposed adaptive normalization. Representative examples are shown in Figure 1, Figure 2 and Figure 3.

3.2.1. Scaling Normalization

Scaling normalization linearly remaps pixel intensities to a fixed range, typically 0–1 after resizing. A representative example is shown in Figure 1. This method is computationally lightweight and commonly used when imaging conditions are relatively homogeneous [1,53].
However, scaling has notable limitations in multi-institutional CXR settings:
  • It does not correct local contrast variations and is sensitive to outliers caused by acquisition artifacts or metallic implants.
  • It fails to harmonize intensity distributions across scanners, leading to degraded cross-domain generalization [16,39].
Accordingly, scaling normalization was included in this study solely as a baseline preprocessing strategy.
Pixel intensities were linearly rescaled from the original 0–255 range to 0–1 following spatial resizing.

3.2.2. Z-Score Normalization

Z-score normalization standardizes image intensities to zero mean and unit variance (Figure 2), reducing brightness and contrast variations caused by heterogeneous imaging devices and patient populations [29,54]. As a result, it is well-suited for multi-center CXR datasets.
A known limitation is its reliance on approximately Gaussian intensity distributions, which are not always present in clinical CXRs [7]. Nevertheless, Z-score normalization consistently outperforms simple scaling in cross-domain evaluations.

3.2.3. Adaptive Normalization (Proposed Method)

The proposed adaptive normalization method addresses both spatial and intensity variability by combining CDF-guided ROI cropping with histogram standardization. This design is motivated by evidence that localized intensity correction improves diagnostic feature visibility and reduces device-induced domain shifts [16,39,52].
A complete visualization of the pipeline is provided in Figure 3.
Step 1: ROI Localization via CDF-Guided Cropping
To reduce background structures that do not contribute to diagnosis, the cumulative distribution function (CDF) of grayscale-sum values is computed:
  • Horizontal (x-axis)—The ROI is extracted between the 5th and 95th percentiles, removing low-density lateral regions that predominantly contain background. This range is selected based on empirical consistency across adult and pediatric CXRs and aligns with findings that lateral regions contribute minimal diagnostic information.
  • Vertical (y-axis)—The ROI is retained between the 15th and 95th percentiles, which excludes anatomical noise above the clavicle and reduces variability caused by neck and shoulder structures.
In exploratory experiments across ChestX-ray14, CheXpert, MIMIC-CXR, and Chest-Xray-Pneumonia, these percentile ranges consistently preserved lung apices and costophrenic angles while excluding most neck and shoulder structures. Percentile-based cropping is also more robust to variations in resolution and aspect ratio than fixed-pixel cropping, which facilitates cross-dataset generalization without manual retuning of crop coordinates.
These percentile thresholds were refined through exploratory analysis and are consistent with established approaches for lung-focused cropping in prior studies [39].
Step 2: Histogram-Based Intensity Standardization
To harmonize contrast across datasets with differing exposure characteristics, image histograms were standardized using target statistics derived from normal ChestX-ray14 images:
  • Target mean: μtarget = 0.4776 × 255 ≈ 121.8;
  • Target standard deviation: σtarget = 0.2238 × 255 ≈ 57.1.
Standardization was applied as defined in Equation (1).
I n o r m x , y = I x , y μ o r i g σ o r i g · σ t a r g e t + μ t a r g e t
where
I x , y : The intensity value of the original image at pixel x , y ;
μ o r i g : Mean intensity of the original image;
σ o r i g : Standard deviation of the original image;
σ t a r g e t : Target standard deviation;
μ t a r g e t : Target mean;
I n o r m x , y : Normalized pixel value at position x , y .
This ensures consistent luminance and contrast across heterogeneous datasets, addressing a key source of domain shift [16].

3.2.4. Summary

Together, these preprocessing techniques ensure that model comparisons are not biased by input variability. While scaling and Z-score normalization provide useful baselines, the proposed adaptive method uniquely addresses both anatomical and intensity-level heterogeneity. By incorporating biologically grounded ROI localization and distribution-aware intensity standardization, the preprocessing pipeline enhances fairness, reproducibility, and cross-domain robustness in clinical deep-learning workflows [52].

3.3. Deep Learning Model Architecture

Three convolutional neural network (CNN) architectures were selected to evaluate how different normalization strategies influence performance and generalization: a custom lightweight CNN, EfficientNet-B0, and MobileNetV2. These models represent a spectrum of computational complexity and capacity, which enables a fair and systematic assessment of preprocessing effects across low, medium, and high expressive architectures. Their selection aligns with widely used CNN-based diagnostic pipelines in chest radiography research [20,42,52].

3.3.1. Custom Lightweight CNN

The custom CNN serves as a controlled baseline for isolating the effect of normalization. Its architecture consists of three convolutional layers with ReLU activation, followed by max-pooling, two fully connected layers, and a Softmax output layer for binary classification. With approximately 1.2 million trainable parameters, the model is intentionally lightweight, enabling rapid training and interpretability. This type of architecture has been recommended for applications in resource-limited settings and embedded diagnostic workflows [21,34]. Figure 4 illustrates the structure of this CNN model.

3.3.2. EfficientNet-B0

EfficientNet-B0 employs compound scaling and squeeze-and-excitation blocks, yielding strong performance with a relatively small parameter footprint. Its robustness in thoracic disease classification and cross-domain generalization has been demonstrated in large-scale radiology studies, making it suitable for evaluating normalization under realistic multi-institutional variation [13,37].

3.3.3. MobileNetV2

MobileNetV2 uses inverted residual blocks with depthwise separable convolutions, optimizing it for low-latency inference on embedded and mobile devices. Despite being lightweight, its performance is highly sensitive to preprocessing quality, particularly when trained on heterogeneous public datasets [5,55,56]. Including MobileNetV2 therefore allows the study to evaluate how normalization affects compact architectures deployed in real-time clinical environments.

3.3.4. Model Training Framework

All models were implemented in PyTorch (version 3.13.5) and trained under identical optimization settings to ensure that performance differences arise solely from the normalization methods. Batch-based stochastic training, cross-entropy loss, and unified tracking of accuracy and F1-score were applied across all architectures. The detailed training workflow is summarized using pseudocode in Section 3.4, replacing the previous code listing in accordance with the Reviewer’s recommendations.

3.4. Experimental Design

The experimental design was structured to evaluate how different normalization strategies affect chest X-ray classification performance under controlled and comparable conditions. An overview of the experimental pipeline is illustrated in Figure 5.
Four publicly available datasets were considered: ChestX-ray14, CheXpert, MIMIC-CXR, and Chest-Xray-Pneumonia. These datasets differ in patient demographics, acquisition protocols, and imaging devices, making them suitable for cross-domain evaluation [17,18,24].
Each dataset was preprocessed using one of three normalization strategies: scaling normalization, Z-score normalization, and adaptive normalization. Three convolutional neural network architectures were evaluated: a custom lightweight CNN, EfficientNet-B0, and MobileNetV2. This resulted in 36 experimental configurations (4 datasets × 3 normalization methods × 3 models).
All experiments were conducted under identical optimization settings to ensure a fair comparison. Patient-level stratified sampling was applied using an 80% training and 20% validation split. To improve robustness, each configuration was evaluated over three independent runs corresponding to different fixed random seeds (42, 123, and 456), and the reported results represent the mean performance across these runs [57,58]. The detailed training procedure is described using pseudocode in Section 3.4.3.

3.4.1. Training Hyperparameters

All models were trained using identical optimization settings. The hyperparameters used in all experiments are summarized in Table 3.
The training workflow is summarized using pseudocode in Section 3.4.3, replacing the previous code listing in accordance with the Reviewer’s recommendations.

3.4.2. Data Augmentation

To avoid confounding effects on intensity normalization, only mild geometric data augmentation was applied uniformly across all experiments. The augmentation strategy included horizontal flipping, small rotations (±7°), isotropic scaling (0.9–1.1), and translation shifts of up to 5% of the image dimensions.
No brightness, contrast, gamma, or intensity jitter was applied to ensure that intensity statistics remained unaffected by augmentation. This strategy follows established recommendations in medical imaging, where geometric transformations improve robustness without altering radiographic intensity patterns [28,46,47].

3.4.3. Training Workflow Pseudocode

The unified training workflow applied to all model and preprocessing combinations is summarized in Algorithm 1. This pseudocode provides a conceptual description of the optimization process and replaces the previous implementation-level code listing, in accordance with the Reviewer’s recommendations.
Algorithm 1. Training Workflow for CXR Classification
Input
Preprocessed training images
Preprocessed validation images
Neural network model M
Hyperparameters from Table 3
Output
Trained model parameters
Procedure
Initialize model M with random weights
For each epoch in the allowed maximum number of epochs
Set model M to training mode
For each batch in the training dataset
  Load batch images and labels
  Perform forward pass to obtain predictions
  Compute cross entropy loss
  Compute gradients through backpropagation
  Update model parameters using the Adam optimizer
End batch loop
Set model M to evaluation mode
Compute accuracy and F1 score on the validation dataset
End epoch loop
Return
The final trained model M
This pseudocode ensures that all models are optimized under identical conditions so that performance differences arise exclusively from the normalization methods.

3.5. Evaluation Metrics and Performance Formulas

3.5.1. Accuracy

Accuracy measures the proportion of correctly classified samples among all predictions. Although widely used, accuracy may become unreliable when dealing with imbalanced medical datasets, which is frequently observed in chest X-ray classification [8]. For this reason, accuracy is reported together with more stable metrics to ensure a balanced evaluation. The formula is provided in Equation (2).
Accuracy = T P + T N T P + F P + T N + F N

3.5.2. F1-Score

The F1-score is the harmonic mean of precision and recall. It provides a balanced perspective by considering both false positives and false negatives. F1-score is particularly valuable in medical imaging because many datasets contain imbalanced class distributions with fewer positive cases [8]. Its definition appears in Equation (3), with the expressions for precision and recall shown in Equation (4).
F 1 = 2 Precision Recall Precision + Recall
where
Precision = T P T P + F P ,                                 Recall = T P T P + F N
In this study, the F1-score is a central evaluation metric, especially for imbalanced datasets such as CheXpert and MIMIC-CXR. To improve reliability, In this study, we report the mean and standard deviation of F1-scores obtained from multiple independent runs using different fixed random seeds.

3.5.3. Sensitivity and Specificity

Sensitivity reflects the proportion of abnormal cases correctly detected by the model. Specificity reflects the proportion of normal cases correctly identified. These measures complement accuracy and F1-score by providing a more detailed analysis of performance under class imbalance, which is crucial for clinical reliability [8]. The formulas are given in Equation (5).
Sensitivity = T P T P + F N ,               Specificity = T N T N + F P
Sensitivity helps evaluate the model’s ability to detect disease, while specificity measures its capability to avoid false alarms. Together, they provide additional insight into the diagnostic behavior of the models across datasets with varying prevalence rates.

3.6. Statistical Significance Testing

To assess whether the performance differences observed across normalization methods and model architectures were statistically meaningful, a three-stage non-parametric testing procedure was applied. This approach follows established recommendations for model comparison in medical image analysis, where performance metrics typically violate assumptions of normality [8,42].
First, the Friedman test was used to examine overall differences across all normalization–model combinations evaluated on multiple datasets. The test is appropriate for repeated-measures settings in which identical classifiers are compared across several conditions without assuming Gaussian distributions [8]. The Friedman test yielded significant results with p-values below 0.05, indicating that at least one method differed from the others.
Following this outcome, the Nemenyi post hoc test was applied to identify which pairs of normalization approaches exhibited statistically significant differences. The Nemenyi test is recommended for pairwise comparisons after Friedman analysis, particularly for experiments involving multiple algorithms evaluated under identical settings [22].
Finally, the Wilcoxon signed-rank test was employed to provide fine-grained paired comparisons of F1-scores between preprocessing methods under identical model architectures and data splits. This test is widely recommended for paired evaluations in cross-validation pipelines and is robust to the non-normal distribution of performance scores [8,59].
The results of the Wilcoxon test are summarized in Table 4.
This statistical analysis further supports the robustness of the observed performance differences, suggesting that the improvements associated with adaptive normalization are unlikely to be attributable to random variation.

4. Experimental Results

This section reports a comprehensive evaluation of the three normalization strategies, scaling, Z-score, and adaptive normalization, across four chest X-ray datasets and three convolutional architectures. Composite figures (Figure 6, Figure 7 and Figure 8) and reorganized performance tables (Table 5, Table 6 and Table 7) were used to summarize the validation accuracy, loss, F1-score, and interaction effects between normalization and model design.

4.1. Accuracy Analysis

Table 5 and Figure 6 summarize the validation accuracy across datasets and model architectures. Across three of the four datasets, adaptive normalization generally maintained or improved accuracy relative to scaling and Z-score normalization, particularly on ChestX-ray14, CheXpert, and Chest-Xray-Pneumonia.
MIMIC-CXR constitutes the main exception, where adaptive normalization yielded slightly lower accuracy than Z-score normalization, reflecting the dataset’s substantial acquisition and intensity heterogeneity.
Overall, the accuracy results indicate that adaptive normalization improves cross-dataset generalization when contrast characteristics are relatively stable, while its effectiveness remains dataset-dependent.
Each block corresponds to one dataset (ChestX-ray14, CheXpert, MIMIC-CXR, Chest X-ray Pneumonia). Bars show the mean validation accuracy over three independent runs for each model (CNN, EfficientNet-B0, MobileNetV2) and normalization strategy (Scaling, Z-score, Adaptive). “Adaptive” refers to the pipeline of CDF-based ROI cropping followed by histogram standardization using the fixed target mean and standard deviation; “Scaling” corresponds to min–max intensity scaling; and “Z-score” uses global dataset-level statistics.

4.2. Loss Analysis

Validation loss results (Table 6 and Figure 7) largely mirror the accuracy trends. In most configurations, adaptive normalization achieved lower or comparable validation loss compared with scaling and Z-score normalization, indicating more stable optimization behavior.
For MIMIC-CXR, adaptive normalization did not provide a consistent loss advantage, again highlighting the challenges posed by strong inter-institution variability.
These findings suggest that adaptive normalization generally stabilizes training dynamics, although its benefits are constrained in datasets with highly heterogeneous acquisition conditions.
Each block corresponds to one dataset (ChestX-ray14, CheXpert, MIMIC-CXR, Chest X-ray Pneumonia). Bars represent the mean validation loss over three runs for each combination of model (CNN, EfficientNet-B0, MobileNetV2) and normalization strategy (Scaling, Z-score, Adaptive). Lower loss values indicate better calibration stability and convergence.

4.3. F1-Score Analysis

Table 7 and Figure 8 report F1-scores across all configurations, providing a balanced evaluation under class imbalance. Adaptive normalization achieved the highest F1-scores on datasets with more consistent intensity characteristics, with the strongest gains observed on Chest-Xray-Pneumonia.
In contrast, Z-score normalization remained competitive or superior on MIMIC-CXR, where broader distributional shifts reduced the effectiveness of histogram-based standardization.
Statistical testing using the Wilcoxon signed-rank test confirmed that adaptive normalization significantly outperformed scaling and Z-score normalization in most dataset–model combinations (p < 0.01), except for MIMIC-CXR.
Each block corresponds to one dataset (ChestX-ray14, CheXpert, MIMIC-CXR, Chest X-ray Pneumonia). Bars represent the mean validation F1-score over three runs for each combination of model (CNN, EfficientNet-B0, MobileNetV2) and normalization strategy (Scaling, Z-score, Adaptive). Higher F1-scores indicate better classification performance and calibration.
Each row shows one test image that was misclassified under Z-score normalization but correctly classified after applying the proposed adaptive normalization. The left column presents the original CXR after Z-score normalization, where low contrast and background dominance contribute to incorrect predictions. The right column shows the same images after CDF-guided cropping and histogram standardization, which enhanced the visibility of lung fields and reduced non-diagnostic background variation. Ground-truth labels and model predictions from MobileNetV2 are indicated in each panel.
Representative examples of misclassified Chest X-ray images before and after normalization are shown in Figure 9.

4.4. Ablation Study: Effect of Cropping and Histogram Standardization

To isolate the contribution of each component in the proposed adaptive normalization pipeline, we conducted an ablation study comparing four conditions: (A) Z-score, (B) cropping only, (C) histogram standardization only, and (D) the full adaptive pipeline. Results on the Chest-Xray-Pneumonia dataset using MobileNetV2 are summarized in Table 8. In this ablation, Z-score normalization serves as the baseline condition, representing input images without either adaptive component (cropping or histogram standardization). For clarity, conditions (A)–(D) correspond directly to the four rows in Table 8.
Histogram standardization produced the largest standalone improvement, increasing the F1-score from 0.85 to 0.88 by harmonizing contrast and suppressing global intensity variability. Cropping yielded a smaller but consistent benefit (0.86) by increasing the lung-to-image ratio and removing non-diagnostic background regions. The full adaptive pipeline achieved the highest performance (0.89), confirming that cropping and histogram standardization are complementary rather than redundant. These results indicate that both components contribute meaningfully to the overall gain, with histogram standardization exerting a stronger individual effect for lightweight architectures such as MobileNetV2.

4.5. Interaction Between Architecture and Normalization

MobileNetV2 exhibited the strongest synergy with adaptive normalization, consistently outperforming or matching the other architectures on datasets with clearer and more uniform intensity characteristics such as ChestX-ray14, CheXpert, and Chest-Xray-Pneumonia. This complementarity arises from several architectural features that are particularly responsive to intensity standardization.
First, depthwise-separable convolutions are highly sensitive to fluctuations in local pixel distribution. Adaptive normalization reduces these fluctuations through region-focused cropping and histogram standardization, leading to more consistent feature activation. Second, the linear bottlenecks in MobileNetV2 benefit from reduced background variation, enabling the model to emphasize diagnostically meaningful structures within the lung fields. The adaptive cropping step further reinforces this effect by focusing the input on anatomically relevant regions.
In datasets where contrast characteristics are relatively stable, these factors collectively support smoother optimization, more stable validation curves, and improved detection of lung opacity patterns. This behavior is also clinically meaningful, as consistent contrast normalization improves the visibility of subtle radiographic abnormalities.
In contrast, for datasets with substantial acquisition variability such as MIMIC-CXR, the advantages of adaptive normalization are less apparent, reflecting the challenges posed by broad cross-institution intensity shifts. Overall, these findings highlight the importance of aligning normalization strategies with model architectures that are particularly receptive to intensity standardization, especially in medical imaging tasks affected by domain variability.

5. Discussion

The experimental results demonstrate that adaptive normalization improves or maintains validation performance relative to scaling and Z-score normalization on three of the four benchmark datasets. This finding is consistent with prior evidence highlighting the importance of intensity normalization and local contrast enhancement in medical image analysis, particularly in multi-center or multi-device settings [7,29,35,39]. By combining percentile-based ROI cropping with histogram standardization, the pipeline reduces non-diagnostic background variation while preserving diagnostically relevant lung structures, leading to more consistent model behavior under moderate domain shifts.
The contrast between the substantial gains observed on ChestX-ray14, CheXpert, and Chest-Xray-Pneumonia and the limited improvements on MIMIC-CXR highlights an important limitation. MIMIC-CXR exhibited pronounced variability in acquisition settings and device characteristics [17,18], weakening the assumptions required for effective histogram-based intensity standardization. Similar observations have been reported in multi-site MRI and radiomics studies, where no single normalization technique consistently dominates under extreme inter-site heterogeneity [7,29,35]. These findings indicate that normalization strategies should be selected according to the expected degree of domain variability rather than applied uniformly across all deployment scenarios. In addition, because pediatric thoraces are proportionally smaller, fixed percentile thresholds may remove relatively larger apical regions, representing a potential source of anatomical bias when adult-derived cropping parameters are applied to pediatric CXRs. This study did not include a dedicated pediatric subgroup analysis; therefore, future work should explicitly validate and, if necessary, re-tune cropping parameters for pediatric populations.
From an architectural perspective, MobileNetV2 combined with adaptive normalization yielded the most reliable performance on datasets with relatively uniform intensity characteristics. This behavior is consistent with the known sensitivity of depthwise-separable convolutions and linear bottlenecks to input distribution consistency [55]. In lightweight architectures, suppressing irrelevant background variation allows representational capacity to focus on subtle parenchymal abnormalities, which is critical for detecting thoracic diseases such as pneumonia and COVID-19 in resource-limited settings [19,20,34]. These results support the view that preprocessing pipelines and model architectures should be co-designed to achieve optimal generalization.
The evaluation framework employed in this study follows established recommendations for rigorous assessment in medical imaging. Reporting F1-scores alongside complementary metrics provides a more reliable evaluation under class imbalance [60]. Statistical significance was assessed using non-parametric tests, including the Friedman–Nemenyi and Wilcoxon signed-rank tests, which are widely endorsed for classifier comparison across multiple datasets [61]. The resulting p-values indicate that the improvements achieved by adaptive normalization were both numerically and statistically meaningful in most scenarios.
Two primary clinical implications emerge from these findings. First, improved robustness under moderate domain shift supports the deployment of CXR-based decision-support systems across institutions with heterogeneous acquisition pipelines, complementing recent advances in efficient and multi-label CXR modeling [6,22,26,45]. Second, the computational efficiency of MobileNetV2, even with the additional preprocessing step, enables near-real-time inference suitable for triage or screening workflows in constrained healthcare environments [34,41].
Adaptive normalization introduced a modest computational overhead. Empirically, the pipeline increased training time by approximately 5.2 ms per batch (+6.3%), with 3.5 ms attributed to CDF-based ROI cropping and 1.7 ms to histogram standardization. This overhead is negligible for offline training and does not affect inference latency because normalization is applied once during preprocessing. Although acceptable in the present setting, such overhead may become relevant for large-scale pretraining or continuous-learning pipelines. Moreover, this study evaluated only four publicly available datasets; performance on unseen clinical cohorts from new scanners, populations, or institutions remains to be systematically assessed. No clinician-in-the-loop assessment or external prospective validation was conducted; therefore, conclusions are limited to retrospective evaluations on publicly available benchmarks.
These limitations motivate several future research directions. Integrating adaptive normalization into federated or source-free domain adaptation frameworks may further improve cross-site robustness while preserving data privacy [14,15]. In addition, combining the pipeline with self-supervised or contrastive pretraining on large-scale unlabeled radiograph repositories could enhance sample efficiency and downstream performance [21,57,58]. Finally, explainability techniques such as Grad-CAM may help elucidate how normalization reshapes salient features, supporting interpretability and clinical adoption [13,56].
Overall, our cross-dataset evaluation reinforces that preprocessing is a critical component of clinically deployable deep learning pipelines. When aligned with dataset characteristics and model architecture, adaptive normalization provides a practical approach for improving generalization in chest X-ray classification while maintaining computational efficiency.

6. Conclusions

This study examined how three preprocessing strategies, scaling, Z-score normalization, and a proposed adaptive normalization pipeline, affect the generalization performance of deep learning models for chest X-ray classification across four benchmark datasets and three convolutional architectures. The findings show that adaptive normalization consistently improves or maintains validation accuracy, loss, and F1-score relative to conventional methods on three of the four datasets, with the strongest performance gains observed when combined with the MobileNetV2 architecture. These improvements reflect more consistent localization of the lung fields and better standardization of intensity distributions, which collectively contribute to more stable optimization and reduced overfitting.
The limited improvements observed on MIMIC-CXR highlight that the effectiveness of normalization depends strongly on the underlying acquisition heterogeneity. In datasets characterized by substantial cross-institution variation, global histogram-based adjustments alone are insufficient, underscoring the need for normalization strategies that explicitly account for broad domain shifts.
The contributions of this study are threefold. First, it provides a cross-dataset, cross-architecture benchmarking framework for systematically evaluating normalization strategies in chest X-ray classification. Second, it offers a statistically grounded comparison using non-parametric tests across multiple metrics, demonstrating that the performance differences are both numerically and statistically meaningful. Third, it demonstrates that an efficient adaptive normalization pipeline integrates well with lightweight architectures such as MobileNetV2, making it suitable for deployment in resource-constrained clinical environments.
Future work will extend this pipeline to more diverse real-world cohorts and explore its integration with federated and source-free domain adaptation frameworks, as well as self-supervised pretraining on large unlabeled CXR collections. Another important direction is to combine adaptive normalization with explainable AI techniques to visualize how preprocessing influences salient regions, thereby facilitating clinical validation and trustworthy deployment of deep learning models in radiographic diagnostics.

Author Contributions

Methodology, J.S. and T.W.; software, J.S.; validation, T.W.; formal analysis, J.S. and T.W.; resources, T.W.; data curation, T.W.; writing—original draft preparation, J.S.; writing—review and editing, J.S. and T.W.; supervision, J.S.; project administration, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study because the data used were obtained from publicly available databases.

Informed Consent Statement

Patient consent was waived because the data used in this study were obtained from publicly available databases.

Data Availability Statement

The datasets used in this study are publicly available from established benchmark repositories (ChestX-ray14, CheXpert, MIMIC-CXR, and Chest X-ray Pneumonia). Due to licensing terms and data governance policies of these repositories, the processed data used in this study are not redistributed publicly but are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Declaration of Generative AI and AI-Assisted Technologies in the Writing Process

During the preparation of this manuscript, the authors used an AI-assisted language tool(ChatGPT, OpenAI, GPT-4) to improve the readability and clarity of the text. After using this tool, the authors carefully reviewed and edited the content and take full responsibility for the integrity, originality, and accuracy of the final manuscript.

References

  1. Oh, Y.; Park, S.; Ye, J.C. Deep Learning COVID-19 Features on CXR Using Limited Training Data Sets. IEEE Trans. Med. Imaging 2020, 39, 2688–2700. [Google Scholar] [CrossRef] [PubMed]
  2. Padmavathi, V.; Ganesan, K. Metaheuristic Optimizers Integrated with Vision Transformer Model for Severity Detection and Classification via Multimodal COVID-19 Images. Sci. Rep. 2025, 15, 13941. [Google Scholar] [CrossRef] [PubMed]
  3. Aksoy, B.; Salman, O.K.M. Detection of COVID-19 Disease in Chest X-Ray Images with Capsul Networks: Application with Cloud Computing. J. Exp. Theor. Artif. Intell. 2021, 33, 527–541. [Google Scholar] [CrossRef]
  4. Khan, A.; Khan, S.H.; Saif, M.; Batool, A.; Sohail, A.; Khan, M.W. A Survey of Deep Learning Techniques for the Analysis of COVID-19 and Their Usability for Detecting Omicron. J. Exp. Theor. Artif. Intell. 2024, 36, 1779–1821. [Google Scholar] [CrossRef]
  5. Marikkar, U.; Atito, S.; Awais, M.; Mahdi, A. LT-ViT: A Vision Transformer for Multi-Label Chest X-Ray Classification. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 2565–2569. [Google Scholar]
  6. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Ball, R.L.; Langlotz, C.; et al. CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning. arXiv 2017, arXiv:1711.05225. [Google Scholar] [CrossRef]
  7. Demircioğlu, A. The Effect of Feature Normalization Methods in Radiomics. Insights Imaging 2024, 15, 2. [Google Scholar] [CrossRef]
  8. Rayed, M.E.; Islam, S.M.S.; Niha, S.I.; Jim, J.R.; Kabir, M.M.; Mridha, M.F. Deep Learning for Medical Image Segmentation: State-of-the-Art Advancements and Challenges. Inform. Med. Unlocked 2024, 47, 101504. [Google Scholar] [CrossRef]
  9. Zhu, Q.; Bai, H.; Sun, J.; Cheng, C.; Li, X. LPAdaIN: Light Progressive Attention Adaptive Instance Normalization Model for Style Transfer. Electronics 2022, 11, 2929. [Google Scholar] [CrossRef]
  10. AbuKaraki, A.; Alrawashdeh, T.; Abusaleh, S.; Alksasbeh, M.Z.; Alqudah, B.; Alemerien, K.; Alshamaseen, H. Pulmonary Edema and Pleural Effusion Detection Using EfficientNet-V1-B4 Architecture and AdamW Optimizer from Chest X-Rays Images. Comput. Mater. Contin. 2024, 80, 1055–1073. [Google Scholar] [CrossRef]
  11. Shin, H.; Kim, T.; Park, J.; Raj, H.; Jabbar, M.S.; Abebaw, Z.D.; Lee, J.; Van, C.C.; Kim, H.; Shin, D. Pulmonary Abnormality Screening on Chest X-Rays from Different Machine Specifications: A Generalized AI-Based Image Manipulation Pipeline. Eur. Radiol. Exp. 2023, 7, 68. [Google Scholar] [CrossRef]
  12. Wangkhamhan, T. Adaptive Chaotic Satin Bowerbird Optimisation Algorithm for Numerical Function Optimisation. J. Exp. Theor. Artif. Intell. 2021, 33, 719–746. [Google Scholar] [CrossRef]
  13. Fan, D.-P.; Zhou, T.; Ji, G.-P.; Zhou, Y.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Inf-Net: Automatic COVID-19 Lung Infection Segmentation From CT Images. IEEE Trans. Med. Imaging 2020, 39, 2626–2637. [Google Scholar] [CrossRef] [PubMed]
  14. Tomar, D.; Lortkipanidze, M.; Vray, G.; Bozorgtabar, B.; Thiran, J.-P. Self-Attentive Spatial Adaptive Normalization for Cross-Modality Domain Adaptation. IEEE Trans. Med. Imaging 2021, 40, 2926–2938. [Google Scholar] [CrossRef] [PubMed]
  15. Luo, Z.; Luo, X.; Gao, Z.; Wang, G. An Uncertainty-Guided Tiered Self-Training Framework for Active Source-Free Domain Adaptation in Prostate Segmentation. arXiv 2024, arXiv:2407.02893. [Google Scholar] [CrossRef]
  16. Gangwar, S.; Devi, R.; Mat Isa, N.A. Optimized Exposer Region-Based Modified Adaptive Histogram Equalization Method for Contrast Enhancement in CXR Imaging. Sci. Rep. 2025, 15, 6693. [Google Scholar] [CrossRef]
  17. Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilcus, S.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.; Shpanskaya, K.; et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI 2019), Honolulu, HI, USA, 27 January–1 February 2019; AAAI Press: Washington, DC, USA, 2019; pp. 590–597. [Google Scholar]
  18. Johnson, A.E.W.; Pollard, T.J.; Berkowitz, S.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.; Mark, R.G.; Horng, S. MIMIC-CXR, a de-Identified Publicly Available Database of Chest Radiographs with Free-Text Reports. Sci. Data 2019, 6, 317. [Google Scholar] [CrossRef]
  19. Rahman, T.; Chowdhury, M.E.H.; Khandakar, A.; Islam, K.R.; Islam, K.F.; Mahbub, Z.B.; Kadir, M.A.; Kashem, S. Transfer Learning with Deep Convolutional Neural Network (CNN) for Pneumonia Detection Using Chest X-Ray. Appl. Sci. 2020, 10, 3233. [Google Scholar] [CrossRef]
  20. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-Ray Images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  21. Gazda, M.; Plavka, J.; Gazda, J.; Drotár, P. Self-Supervised Deep Convolutional Neural Network for Chest X-Ray Classification. IEEE Access 2021, 9, 151972–151982. [Google Scholar] [CrossRef]
  22. Öztürk, Ş.; Turalı, M.Y.; Çukur, T. HydraViT: Adaptive Multi-Branch Transformer for Multi-Label Disease Classification from Chest X-Ray Images. Biomed. Signal Process. Control. 2025, 100, 106959. [Google Scholar] [CrossRef]
  23. Dede, A.; Nunoo-Mensah, H.; Tchao, E.T.; Agbemenu, A.S.; Adjei, P.E.; Acheampong, F.A.; Kponyo, J.J. Deep Learning for Efficient High-Resolution Image Processing: A Systematic Review. Intell. Syst. Appl. 2025, 26, 200505. [Google Scholar] [CrossRef]
  24. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell 2018, 172, 1122.e9–1131.e9. [Google Scholar] [CrossRef]
  25. Ahmad, I.S.; Li, N.; Wang, T.; Liu, X.; Dai, J.; Chan, Y.; Liu, H.; Zhu, J.; Kong, W.; Lu, Z.; et al. COVID-19 Detection via Ultra-Low-Dose X-Ray Images Enabled by Deep Learning. Bioengineering 2023, 10, 1314. [Google Scholar] [CrossRef] [PubMed]
  26. Oltu, B.; Güney, S.; Yuksel, S.E.; Dengiz, B. Automated Classification of Chest X-Rays: A Deep Learning Approach with Attention Mechanisms. BMC Med. Imaging 2025, 25, 71. [Google Scholar] [CrossRef] [PubMed]
  27. Saad, M.M.; Rehmani, M.H.; O’Reilly, R. Addressing the Intra-Class Mode Collapse Problem Using Adaptive Input Image Normalization in GAN-Based X-Ray Images. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, Scotland, 11–15 July 2022; pp. 2049–2052. [Google Scholar]
  28. Islam, T.; Hafiz, M.S.; Jim, J.R.; Kabir, M.M.; Mridha, M.F. A Systematic Review of Deep Learning Data Augmentation in Medical Imaging: Recent Advances and Future Research Directions. Healthc. Anal. 2024, 5, 100340. [Google Scholar] [CrossRef]
  29. Reinhold, J.C.; Dewey, B.E.; Carass, A.; Prince, J.L. Evaluating the Impact of Intensity Normalization on MR Image Synthesis. In Proceedings of the Medical Imaging 2019: Image Processing, San Diego, CA, USA, 19–21 February 2019; Angelini, E.D., Landman, B.A., Eds.; SPIE: Bellingham, WA, USA, 2019; Volume 10949, p. 109493H. [Google Scholar]
  30. Huang, X.; Belongie, S. Arbitrary Style Transfer in Real-Time with Adaptive Instance Normalization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1510–1519. [Google Scholar]
  31. Al-Waisy, A.S.; Mohammed, M.A.; Al-Fahdawi, S.; Maashi, M.S.; Garcia-Zapirain, B.; Abdulkareem, K.H.; Mostafa, S.A.; Kumar, N.M.; Le, D.-N. COVID-DeepNet: Hybrid Multimodal Deep Learning System for Improving COVID-19 Pneumonia Detection in Chest X-Ray Images. Comput. Mater. Contin. 2021, 67, 2409–2429. [Google Scholar] [CrossRef]
  32. bin Azhar, A.N.; Sani, N.S.; Wei, L.L.X. Enhancing COVID-19 Detection in X-Ray Images Through Deep Learning Models with Different Image Preprocessing Techniques. Int. J. Adv. Comput. Sci. Appl. 2025, 16, 633–644. [Google Scholar] [CrossRef]
  33. Toğaçar, M.; Ergen, B.; Cömert, Z. COVID-19 Detection Using Deep Learning Models to Exploit Social Mimic Optimization and Structured Chest X-Ray Images Using Fuzzy Color and Stacking Approaches. Comput. Biol. Med. 2020, 121, 103805. [Google Scholar] [CrossRef]
  34. Sanida, T.; Dasygenis, M. A Novel Lightweight CNN for Chest X-Ray-Based Lung Disease Identification on Heterogeneous Embedded System. Appl. Intell. 2024, 54, 4756–4780. [Google Scholar] [CrossRef]
  35. Albert, S.; Wichtmann, B.D.; Zhao, W.; Maurer, A.; Hesser, J.; Attenberger, U.I.; Schad, L.R.; Zöllner, F.G. Comparison of Image Normalization Methods for Multi-Site Deep Learning. Appl. Sci. 2023, 13, 8923. [Google Scholar] [CrossRef]
  36. Bani Baker, Q.; Hammad, M.; Al-Smadi, M.; Al-Jarrah, H.; Al-Hamouri, R.; Al-Zboon, S.A. Enhanced COVID-19 Detection from X-Ray Images with Convolutional Neural Network and Transfer Learning. J. Imaging 2024, 10, 250. [Google Scholar] [CrossRef] [PubMed]
  37. Tan, M.; Le, Q. V EfficientNetV2: Smaller Models and Faster Training. arXiv 2011, arXiv:2104.00298. [Google Scholar] [CrossRef]
  38. Pavlova, M.; Terhljan, N.; Chung, A.G.; Zhao, A.; Surana, S.; Aboutalebi, H.; Gunraj, H.; Sabri, A.; Alaref, A.; Wong, A. COVID-Net CXR-2: An Enhanced Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-Ray Images. Front. Med. 2022, 9, 861680. [Google Scholar] [CrossRef] [PubMed]
  39. Philipsen, R.H.H.M.; Maduskar, P.; Hogeweg, L.; Melendez, J.; Sánchez, C.I.; van Ginneken, B. Localized Energy-Based Normalization of Medical Images: Application to Chest Radiography. IEEE Trans. Med. Imaging 2015, 34, 1965–1975. [Google Scholar] [CrossRef]
  40. Gunraj, H.; Wang, L.; Wong, A. COVIDNet-CT: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases From Chest CT Images. Front. Med. 2020, 7, 608525. [Google Scholar] [CrossRef]
  41. Sriwiboon, N. Efficient and Lightweight CNN Model for COVID-19 Diagnosis from CT and X-Ray Images Using Customized Pruning and Quantization Techniques. Neural Comput. Appl. 2025, 37, 13059–13078. [Google Scholar] [CrossRef]
  42. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Rajendra Acharya, U. Automated Detection of COVID-19 Cases Using Deep Neural Networks with X-Ray Images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef]
  43. Yen, C.-T.; Tsao, C.-Y. Lightweight Convolutional Neural Network for Chest X-Ray Images Classification. Sci. Rep. 2024, 14, 29759. [Google Scholar] [CrossRef]
  44. Hage Chehade, A.; Abdallah, N.; Marion, J.-M.; Hatt, M.; Oueidat, M.; Chauvet, P. Reconstruction-Based Approach for Chest X-Ray Image Segmentation and Enhanced Multi-Label Chest Disease Classification. Artif. Intell. Med. 2025, 165, 103135. [Google Scholar] [CrossRef]
  45. Fu, X.; Lin, R.; Du, W.; Tavares, A.; Liang, Y. Explainable Hybrid Transformer for Multi-Classification of Lung Disease Using Chest X-Rays. Sci. Rep. 2025, 15, 6650. [Google Scholar] [CrossRef]
  46. Rahman, M.; Gao, J.; Carey, K.A.; Edelson, D.P.; Afshar, A.; Garrett, J.W.; Chen, G.; Afshar, M.; Churpek, M.M. Comparison of Deep Learning Approaches Using Chest Radiographs for Predicting Clinical Deterioration: Retrospective Observational Study. JMIR AI 2025, 4, e67144. [Google Scholar] [CrossRef] [PubMed]
  47. Mirugwe, A.; Tamale, L.; Nyirenda, J. Improving Tuberculosis Detection in Chest X-Ray Images Through Transfer Learning and Deep Learning: Comparative Study of Convolutional Neural Network Architectures. JMIRx Med 2025, 6, e66029. [Google Scholar] [CrossRef] [PubMed]
  48. Boecking, B.; Usuyama, N.; Bannur, S.; Castro, D.C.; Schwaighofer, A.; Hyland, S.; Wetscherek, M.; Naumann, T.; Nori, A.; Alvarez-Valle, J.; et al. Making the Most of Text Semantics to Improve Biomedical Vision–Language Processing. In Proceedings of the ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 1–21. [Google Scholar]
  49. Ma, D.; Pang, J.; Gotway, M.B.; Liang, J. A Fully Open AI Foundation Model Applied to Chest Radiography. Nature 2025, 643, 488–498. [Google Scholar] [CrossRef] [PubMed]
  50. Wang, Z.; Wu, Z.; Agarwal, D.; Sun, J. MedCLIP: Contrastive Learning from Unpaired Medical Images and Text. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; Goldberg, Y., Kozareva, Z., Zhang, Y., Eds.; Association for Computational Linguistics: Abu Dhabi, United Arab Emirates, 2022; pp. 3876–3887. [Google Scholar]
  51. Zhang, Y.; Jiang, H.; Miura, Y.; Manning, C.D.; Langlotz, C. Contrastive Learning of Medical Visual Representations from Paired Images and Text. In Proceedings of the Machine Learning in Health Care, Virtual, 7–8 August 2020. [Google Scholar]
  52. Iqbal, H.; Khan, A.; Nepal, N.; Khan, F.; Moon, Y.-K. Deep Learning Approaches for Chest Radiograph Interpretation: A Systematic Review. Electronics 2024, 13, 4688. [Google Scholar] [CrossRef]
  53. Chowdhury, N.K.; Rahman, M.M.; Kabir, M.A. PDCOVIDNet: A Parallel-Dilated Convolutional Neural Network Architecture for Detecting COVID-19 from Chest X-Ray Images. Health Inf. Sci. Syst. 2020, 8, 27. [Google Scholar] [CrossRef]
  54. Banik, P.; Majumder, R.; Mandal, A.; Dey, S.; Mandal, M. A Computational Study to Assess the Polymorphic Landscape of Matrix Metalloproteinase 3 Promoter and Its Effects on Transcriptional Activity. Comput. Biol. Med. 2022, 145, 105404. [Google Scholar] [CrossRef]
  55. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  56. Bougourzi, F.; Dornaika, F.; Distante, C.; Taleb-Ahmed, A. D-TrAttUnet: Toward Hybrid CNN-Transformer Architecture for Generic and Subtle Segmentation in Medical Images. Comput. Biol. Med. 2024, 176, 108590. [Google Scholar] [CrossRef]
  57. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  58. Zhou, Z.; Sodha, V.; Pang, J.; Gotway, M.B.; Liang, J. Models Genesis. Med. Image Anal. 2021, 67, 101840. [Google Scholar] [CrossRef]
  59. Yan, Z.; Li, X.; Li, M.; Zuo, W.; Shan, S. Shift-Net: Image Inpainting via Deep Feature Rearrangement. In Proceedings of the European Conference on Computer Vision (ECCV 2018), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Matsushita, Y., Eds.; Springer: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  60. Chicco, D.; Jurman, G. The Advantages of the Matthews Correlation Coefficient (MCC) over F1 Score and Accuracy in Binary Classification Evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef]
  61. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. Available online: https://www.jmlr.org/papers/v7/demsar06a.html (accessed on 27 July 2025).
Figure 1. Example of scaling normalization applied after image resizing. The dashed line represents the linear mapping from pixel intensity values (0–255) to the normalized range (0–1), and the dots indicate representative pixel values before and after scaling.
Figure 1. Example of scaling normalization applied after image resizing. The dashed line represents the linear mapping from pixel intensity values (0–255) to the normalized range (0–1), and the dots indicate representative pixel values before and after scaling.
Jimaging 12 00014 g001
Figure 2. Example of Z-score normalization applied after image resizing. The dashed line represents the linear transformation from pixel intensity values (0–255) to standardized Z-score values, while the dots indicate representative pixel values before and after normalization.
Figure 2. Example of Z-score normalization applied after image resizing. The dashed line represents the linear transformation from pixel intensity values (0–255) to standardized Z-score values, while the dots indicate representative pixel values before and after normalization.
Jimaging 12 00014 g002
Figure 3. Visualization of the adaptive normalization pipeline. (1) Original CXR image, (2) grayscale-sum profiles computed along the horizontal (column-wise) and vertical (row-wise) axes, (3) CDF-based percentile thresholds, where the solid colored lines denote the selected lower and upper cutoffs used for ROI cropping, (4) grayscale histograms before and after normalization, (5) cropped ROI image, and (6) final normalized image.
Figure 3. Visualization of the adaptive normalization pipeline. (1) Original CXR image, (2) grayscale-sum profiles computed along the horizontal (column-wise) and vertical (row-wise) axes, (3) CDF-based percentile thresholds, where the solid colored lines denote the selected lower and upper cutoffs used for ROI cropping, (4) grayscale histograms before and after normalization, (5) cropped ROI image, and (6) final normalized image.
Jimaging 12 00014 g003
Figure 4. Conceptual representation of the lightweight CNN used in this study. The arrows indicate the sequential data flow through the network layers, while the different colors denote distinct functional stages, including feature extraction, feature refinement, spatial downsampling, and classification.
Figure 4. Conceptual representation of the lightweight CNN used in this study. The arrows indicate the sequential data flow through the network layers, while the different colors denote distinct functional stages, including feature extraction, feature refinement, spatial downsampling, and classification.
Jimaging 12 00014 g004
Figure 5. Overview of the experimental pipeline illustrating datasets, normalization strategies, model architectures, and binary classification task. The arrows indicate the sequential flow of data through the experimental pipeline, while the different colors are used to distinguish major stages, including dataset sources, image preprocessing techniques, deep learning model architectures, and the final classification task.
Figure 5. Overview of the experimental pipeline illustrating datasets, normalization strategies, model architectures, and binary classification task. The arrows indicate the sequential flow of data through the experimental pipeline, while the different colors are used to distinguish major stages, including dataset sources, image preprocessing techniques, deep learning model architectures, and the final classification task.
Jimaging 12 00014 g005
Figure 6. Composite validation accuracy across datasets and model architectures.
Figure 6. Composite validation accuracy across datasets and model architectures.
Jimaging 12 00014 g006
Figure 7. Composite validation loss across datasets and model architectures.
Figure 7. Composite validation loss across datasets and model architectures.
Jimaging 12 00014 g007
Figure 8. Composite validation F1-score across datasets and model architectures.
Figure 8. Composite validation F1-score across datasets and model architectures.
Jimaging 12 00014 g008
Figure 9. Misclassified Chest-Xray-Pneumonia cases before and after adaptive normalization. Examples show cases incorrectly classified under Z-score normalization but correctly classified after applying the proposed adaptive normalization.
Figure 9. Misclassified Chest-Xray-Pneumonia cases before and after adaptive normalization. Examples show cases incorrectly classified under Z-score normalization but correctly classified after applying the proposed adaptive normalization.
Jimaging 12 00014 g009
Table 1. Comparative Summary of Deep Learning, Normalization, and Augmentation Studies in Chest X-ray/CT Imaging (2020–2025). Reorganized into three groups as requested by the Reviewers: (A) Deep Learning Architectures for CXR/CT Classification, (B) Normalization and Preprocessing Approaches, and (C) Data Augmentation and Transferability Studies.
Table 1. Comparative Summary of Deep Learning, Normalization, and Augmentation Studies in Chest X-ray/CT Imaging (2020–2025). Reorganized into three groups as requested by the Reviewers: (A) Deep Learning Architectures for CXR/CT Classification, (B) Normalization and Preprocessing Approaches, and (C) Data Augmentation and Transferability Studies.
A. Deep Learning Architectures for CXR/CT Classification.
AuthorsModel/ArchitectureTechnique/ApproachDatasetMetricsKey Highlights
[13]Inf-NetSemi-supervised infection segmentation with reverse and edge attentionCOVID-SemiSegDice, Sens., Spec.First semi-supervised CT infection segmentation dataset and model
[1]ResNet-18 + CAMPatch-based semi-supervised CXR learningCOVIDx, RSNAAccuracy, AUCStrong performance with limited labeled CXRs
[19]VGG19, DenseNet, InceptionTransfer learning with augmentationChest X-ray PneumoniaAccuracy, F1Demonstrated impact of augmentation and TL
[40]COVIDNet-CTCOVID-specific CNNCOVIDx-CTAccuracy, Sens.High-performing CT-based classifier
[42]DarkCOVIDNetCNN for binary/multi-class COVID-19 detectionCOVID-19 X-rayAccuracy, F1Early high-performing CXR classifier
[33]MobileNetV2 + SqueezeNetFuzzy preprocessing with metaheuristic optimizationCOVID (Cohen)Accuracy, F1Fusion of deep and fuzzy features
[20]COVID-NetMachine-designed CNNCOVIDxAccuracy, Sens.Transparent, explainable CNN design
[31]InceptionResNetV2 + BiLSTMHybrid deep features (GLCM/LBP + CNN)COVIDxAccuracy, AUCOutperformed CNN-only baselines
[5]LT-ViTLabel-token vision transformerCheXpert, CXR14AUCInterpretable ViT via explicit label tokens
[11]ResNet-basedStyle and histogram normalization pipelineMulti-hospital CXRAUC, AccuracyCross-device robust AI pipeline
[10]EfficientNet-B4CLAHE + augmentationCXR14, PadChestF1, AUCHigh robustness with strong AUC
[34]Lightweight CNNEmbedded-optimized CNNCXR14Accuracy, Sens.Real-time classification on heterogeneous devices
[43]Lightweight CNNEdge-oriented deploymentCXR14Accuracy, Precision, Sens.Ultra-light model with near-ResNet performance
[44]Reconstruction-based CNN + XGBoostGAN reconstruction, segmentation + radiomicsCXR14AUC, AccuracyPathology-aware reconstruction and classification
[45]EHTNet (Hybrid CNN–Transformer)Explainable hybrid transformer for lung diseasesCXR14Accuracy, AUCHybrid CNN–ViT architecture with attention-based explanations
[2]ViT + MetaheuristicsViT optimized with PSO/GWO for severity and multimodal fusionCXR + CTAccuracy, Sens.ViT tuned via metaheuristic optimizers
This StudyCNN, EfficientNet-B0, MobileNetV2Adaptive normalization (CDF cropping + histogram standardization)CXR14, CheXpert, MIMIC-CXR, PediatricAccuracy, F1A systematic cross-dataset, cross-architecture benchmarking of normalization strategies under controlled sampling.
B. Normalization and Preprocessing Approaches
AuthorsMethodDatasetMetricsKey Highlights
[39]Localized energy-based normalizationChest radiographySens., Spec.Foundational localized contrast standardization for CXRs
[30]Adaptive Instance Normalization (AdaIN)Natural + style datasetsStyle metricsBasis for modern adaptive, style-aware normalization
[14]Self-attentive spatial adaptive normalizationRadiography/CTDice, IoUSpatially adaptive normalization for cross-modality DA
[11]Histogram + style normalization pipelineMulti-hospital CXRAccuracy, AUCReduced device-driven variation across machines
[16]Exposure-region adaptive histogram equalizationCXRPSNR, entropyModern exposure-region contrast enhancement method
This StudyCDF-guided cropping + histogram standardizationCXR14, CheXpert, MIMIC-CXRAccuracy, F1Joint ROI and intensity standardization for multi-source CXRs
C. Data Augmentation and Transferability Studies
AuthorsFocusDataset(s)MetricsKey Highlights
[28]Systematic review of data augmentation in medical imagingMulti-modal medicalNarrativeIdentified geometric transforms as backbone of medical DL augmentation
[46]Augmentation for clinical deterioration predictionCXRAccuracy, AUCShowed rotation and flipping improve robustness in clinical prediction
[47]Augmentation for TB/COVID robustnessMultiple CXR datasetsAUC, PrecisionBrightness and gamma augmentations improve cross-dataset transferability
[33]Fuzzy preprocessing + augmentationCXR (COVID)Accuracy, F1Demonstrated that fuzzy preprocessing with augmentation enhances performance
Table 2. Summary of Chest X-ray Datasets Used in Normalization Evaluation.
Table 2. Summary of Chest X-ray Datasets Used in Normalization Evaluation.
DatasetImage CountImages UsedPatientsOriginal
Labels/Classes
Classes Used
ChestX-ray14112,12016,00030,80514 2
CheXpert224,31616,00065,24014 2
MIMIC-CXR377,11016,000227,82714 2
Chest-Xray-Pneumonia58635863Pediatric only3 2
Table 3. Hyperparameters used for all experiments.
Table 3. Hyperparameters used for all experiments.
HyperparameterValue
OptimizerAdam
Learning rate1 × 10−4
Batch size100
Maximum epochs20
Weight decay1 × 10−5
Train validation split80%/20%
Random seeds42, 123, 456
Loss functionCross entropy
Table 4. Wilcoxon Signed-Rank Test Results (n = 12 paired dataset × model configurations).
Table 4. Wilcoxon Signed-Rank Test Results (n = 12 paired dataset × model configurations).
Comparisonp-ValueSignificance
Adaptive vs. Z-score0.0078Significant (p < 0.01)
Adaptive vs. Scaling0.0039Significant (p < 0.01)
Z-score vs. Scaling0.0781Not significant (p > 0.05)
Table 5. Summary of Validation Accuracy Across Datasets and Configurations.
Table 5. Summary of Validation Accuracy Across Datasets and Configurations.
DatasetDeep LearningScalingZ-ScoreAdaptive
ChestX-ray14CNN0.580.590.59
EfficientNet-B00.620.610.62
MobileNetV20.640.620.64
CheXpertCNN0.840.850.85
EfficientNet-B00.850.840.86
MobileNetV20.810.840.83
MIMIC-CXRCNN0.650.640.64
EfficientNet-B00.650.650.60
MobileNetV20.620.630.60
Chest-Xray-PneumoniaCNN0.610.630.63
EfficientNet-B00.830.870.82
MobileNetV20.820.880.91
Cells highlighted in green indicate the highest validation accuracy within each dataset and model configuration. “Adaptive” refers to the proposed adaptive normalization pipeline consisting of CDF-based ROI cropping followed by histogram standardization using fixed target mean and standard deviation. “Scaling” corresponds to min–max scaling to the [0, 1] range, while “Z-score” normalization uses the global dataset-level mean and standard deviation. All results are averaged over three runs with fixed random seeds and identical stratified data splits.
Table 6. Summary of Validation Loss Across Datasets and Configurations.
Table 6. Summary of Validation Loss Across Datasets and Configurations.
DatasetPreprocessingScalingZ-ScoreAdaptive
ChestX-ray14CNN0.680.670.69
EfficientNet-B00.660.670.68
MobileNetV20.660.730.71
CheXpertCNN0.470.460.46
EfficientNet-B00.420.400.44
MobileNetV20.430.390.37
MIMIC-CXRCNN0.660.650.65
EfficientNet-B00.700.660.68
MobileNetV20.700.690.74
Chest-Xray-PneumoniaCNN0.720.700.70
EfficientNet-B00.360.400.40
MobileNetV20.450.400.29
Lower loss values indicate better model calibration and convergence. All loss values represent averages over three experimental runs.
Table 7. Summary of Validation F1-scores Across Datasets and Configurations.
Table 7. Summary of Validation F1-scores Across Datasets and Configurations.
DatasetPreprocessingScalingZ-ScoreAdaptive
ChestX-ray14CNN0.570.580.59
EfficientNet-B00.620.600.60
MobileNetV20.640.620.62
CheXpertCNN0.840.850.85
EfficientNet-B00.840.840.84
MobileNetV20.790.830.82
MIMIC-CXRCNN0.650.640.64
EfficientNet-B00.640.640.60
MobileNetV20.620.630.60
Chest-Xray-PneumoniaCNN0.610.630.63
EfficientNet-B00.810.840.81
MobileNetV20.800.850.89
Higher F1-scores indicate better model calibration and classification performance. All values represent averages over three experimental runs.
Table 8. Ablation Study Results (MobileNetV2, Chest-Xray-Pneumonia).
Table 8. Ablation Study Results (MobileNetV2, Chest-Xray-Pneumonia).
MethodCroppingHistogramF1-Score
Z-score0.85
Cropping only0.86
Histogram only0.88
Adaptive0.89
All values represent the mean of three runs using identical stratified splits and the hyperparameters described in Section 3.4. ✓ indicates that the corresponding component is applied, while ✗ indicates that it is not applied.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Singthongchai, J.; Wangkhamhan, T. Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification. J. Imaging 2026, 12, 14. https://doi.org/10.3390/jimaging12010014

AMA Style

Singthongchai J, Wangkhamhan T. Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification. Journal of Imaging. 2026; 12(1):14. https://doi.org/10.3390/jimaging12010014

Chicago/Turabian Style

Singthongchai, Jatsada, and Tanachapong Wangkhamhan. 2026. "Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification" Journal of Imaging 12, no. 1: 14. https://doi.org/10.3390/jimaging12010014

APA Style

Singthongchai, J., & Wangkhamhan, T. (2026). Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification. Journal of Imaging, 12(1), 14. https://doi.org/10.3390/jimaging12010014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop