Next Article in Journal
Lung Nodule Malignancy Classification Integrating Deep and Radiomic Features in a Three-Way Attention-Based Fusion Module
Previous Article in Journal
Lightweight Statistical and Texture Feature Approach for Breast Thermogram Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Non-Contrast Brain CT Images Segmentation Enhancement: Lightweight Pre-Processing Model for Ultra-Early Ischemic Lesion Recognition and Segmentation

1
Higher School of Digital Culture, ITMO University, St. Petersburg 197101, Russia
2
Faculty of Radio Engineering and Telecommunications, St. Petersburg Electrotechnical University “LETI”, St. Petersburg 197022, Russia
3
Department of Microbiological Synthesis Technology, St. Petersburg State Institute of Technology, St. Petersburg 190013, Russia
4
Mathematics and Mechanics Faculty, St. Petersburg State University, St. Petersburg 198504, Russia
5
Information Systems Department, International IT University, Almaty 050000, Kazakhstan
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(10), 359; https://doi.org/10.3390/jimaging11100359 (registering DOI)
Submission received: 22 August 2025 / Revised: 19 September 2025 / Accepted: 8 October 2025 / Published: 13 October 2025
(This article belongs to the Section Medical Imaging)

Abstract

Timely identification and accurate delineation of ultra-early ischemic stroke lesions in non-contrast computed tomography (CT) scans of the human brain are of paramount importance for prompt medical intervention and improved patient outcomes. In this study, we propose a deep learning-driven methodology specifically designed for segmenting ultra-early ischemic regions, with a particular emphasis on both the ischemic core and the surrounding penumbra during the initial stages of stroke progression. We introduce a lightweight preprocessing model based on convolutional filtering techniques, which enhances image clarity while preserving the structural integrity of medical scans, a critical factor when detecting subtle signs of ultra-early ischemic strokes. Unlike conventional preprocessing methods that directly modify the image and may introduce artifacts or distortions, our approach ensures the absence of neural network-induced artifacts, which is especially crucial for accurate diagnosis and segmentation of ultra-early ischemic lesions. The model employs predefined differentiable filters with trainable parameters, allowing for artifact-free and precision-enhanced image refinement tailored to the challenges of ultra-early stroke detection. In addition, we incorporated into the combined preprocessing pipeline a newly proposed trainable linear combination of pretrained image filters, a concept first introduced in this study. For model training and evaluation, we utilize a publicly available dataset of acute ischemic stroke cases, focusing on the subset relevant to ultra-early stroke manifestations, which contains annotated non-contrast CT brain scans from 112 patients. The proposed model demonstrates high segmentation accuracy for ultra-early ischemic regions, surpassing existing methodologies across key performance metrics. The results have been rigorously validated on test subsets from the dataset, confirming the effectiveness of our approach in supporting the early-stage diagnosis and treatment planning for ultra-early ischemic strokes.

1. Introduction

In the domain of contemporary medicine, strokes and other consequences of ischemic attacks represent one of the primary causes of mortality and disability worldwide. According to the World Health Organization (WHO), approximately 15 million individuals suffer from strokes annually, with 5 million resulting in fatal outcomes and another 5 million leading to long-term disability [1]. The increasing prevalence of ischemic strokes is closely correlated with the rise in risk factors such as hypertension, diabetes mellitus, obesity, and sedentary lifestyles. Despite advancements in medical technologies and treatment methodologies, the long-term repercussions of strokes continue to exert a profound impact on public health and the quality of life for millions globally [2].
The prevention of strokes and the early detection of ischemic attacks constitute critical aspects of modern medical diagnostics. Magnetic Resonance Imaging (MRI) [3] plays a pivotal role in this process, providing high-resolution cross-sectional images of the brain, which enable the identification and quantification of ischemic regions. Additionally, non-contrast computed tomography (NCCT) [4] is widely employed. Early stroke detection is particularly crucial, as it facilitates the precise delineation of ischemic zones and initiates immediate medical intervention and treatment. However, the manual interpretation of medical imaging by specialists presents a formidable challenge due to the subtle and intricate nature of ischemic lesions [5], which are often difficult to discern with the naked eye. This underscores the necessity of leveraging advanced computational methodologies to accurately identify and analyze affected regions.
To address these challenges, modern technologies, particularly deep neural networks (DNNs) [6,7,8,9,10,11,12,13,14,15,16,17,18,19], have emerged as a promising solution. These models have demonstrated high performance in solving computer vision tasks, including image segmentation, which is crucial in automating the detection of ischemic regions. Image segmentation involves mapping a pixel-wise mask of the lesion area onto the original medical image, thereby automating the detection process and significantly improving accuracy and efficiency.
In this study, we focus on the task of segmenting ischemic regions in non-contrast computed tomography (NCCT) scans [4]. Our approach addresses two critical segmentation tasks:
  • Segmentation of the Ischemic Penumbra [20]. This region of the ischemic lesion includes brain tissues and cells that have the potential to recover with medical intervention.
  • Segmentation of the Ischemic Core [21]. This area consists of cells that have irreversibly lost their function.
By accurately segmenting both the penumbra and the core, our method provides a comprehensive analysis of the ischemic regions, which is crucial for effective treatment planning. Automating ischemia region segmentation requires high-quality segmentation models. Furthermore, in industrial settings, cost reduction remains a critical factor, necessitating the development of computationally efficient models.
Therefore, we focus on developing a lightweight image preprocessing approach to enhance the segmentation quality of the ischemic penumbra and ischemic core region segmentation. Preprocessing models have recently been widely adopted in medical applications and have proven their effectiveness [22,23]. The proposed model effectively mitigates artifacts such as low contrast, high/low brightness, blurring, and noise, thereby improving image quality and, consequently, the accuracy of image segmentation models. Artifacts in CT scans [24,25] arise from various factors, including the limited contrast range of soft tissues, patient motion artifacts, interpolation artifacts, and other hardware and software limitations of this imaging technique [26]. The use of contrast agents can improve diagnostic accuracy and partially mitigate artifacts; however, this method is more expensive, has several contraindications, and requires a prolonged acquisition time, which is unacceptable in cases of ischemic stroke, where every second is critical.
The development of generative computer vision (including GAN-based models, diffusion models, and autoencoders) has also influenced the task of improving image color enhancement. Existing state-of-the-art solutions are typically end-to-end architectures [27,28,29,30,31] and tend to be computationally heavy. The direct application of generative approaches without additional processing is unsuitable for medical diagnostic tasks, as such models are prone to generating various neural network artifacts and may fabricate details that are not present in the original image.
A promising approach to improving image quality is the use of combined neural network methods based on filters [29,32,33,34,35,36], which integrate neural networks for parameter generation with classical mathematical differentiable functions for modeling various filters, such as contrast adjustment, sharpening, white balance correction, and others.
Our proposed model employs a combined neural network approach that addresses these limitations. It is more robust and incapable of generating artifacts, as it does not directly modify the image but instead predicts parameters for predefined differentiable filters.
We introduce a lightweight combined image preprocessing model for ischemia region segmentation. To evaluate the effectiveness of the proposed preprocessing model, we conducted a series of experiments that compared the quality of the segmentation of ischemic regions before and after applying our approach. We used several key metrics, such as IoU 3D and Dice 3D, to assess the impact of preprocessing on segmentation outcomes objectively. As a result, our model demonstrated a significant improvement in segmentation accuracy.
Within the scope of our study, we developed a novel lightweight, trainable neural network block composed of image filters that enhance the effectiveness of biomedical image processing tasks within the preprocessing pipeline proposed in this work. The suggested approach is based on a learnable linear combination of convolutions with varying kernel sizes, employing pretrained filter kernels. These kernels were trained using an encoder-decoder framework on a subset of the employed dataset. This innovation enables improved performance of the overall preprocessing pipeline, offering an efficient solution for enhancing image quality and feature representation in biomedical imaging applications.
Our approach significantly enhances the accuracy of identifying critical health issues in patients. It facilitates precise and timely intervention for managing the consequences of ischemic attacks and enables the implementation of preventive measures to avert strokes and other adverse outcomes. The improved recognition and segmentation capabilities contribute to better patient outcomes and more effective healthcare interventions, potentially reducing the incidence and severity of stroke-related disabilities.
The contributions of this work can be summarized as follows:
  • We propose a lightweight combined neural network approach for preprocessing non-contrast computer tomography (NCCT) brain images to enhance the segmentation quality of the ischemic core and penumbra regions.
  • We also introduce a lightweight, trainable image filtering module based on a learnable linear combination of convolutions with pretrained kernels of multiple sizes, integrated into a biomedical image preprocessing pipeline to enhance task performance.
  • As the baseline architecture for the preprocessing model, we use the LFIEM/UNIFI model [32,33], with minor modifications to its architecture and image processing filters.
  • Our approach demonstrates improved performance on the CPAISD dataset [37] (extemely hard case), combined RSNA-MICCAI [38] and CQ500 [39] (as a general case) in terms of IoU 3D and Dice 3D metrics using various segmentation pipelines.

2. Problem Statement

This study primarily focuses on enhancing the performance of multiclass segmentation in brain CT images. The input image is segmented into three regions: ischemic core, penumbra area, and background. The segmentation model produces three output masks, each corresponding to a specific region, with the label “1” representing a significant region and “0” otherwise (Figure 1).

3. Materials and Methods

The conceptually developed solution is a two-stage process and consists of an Image Preprocessing Module and an Image Segmentation Module, as shown in Figure 2. The preprocessing block is necessary to improve the original image, while the segmentation block follows to segment the image and generate the prediction masks, respectively. In the following subsections, we discuss each block in more detail and the specifics of its joint training.

3.1. Image Preprocessing Module

Our methodology is based on the theoretical framework outlined in the reference works [32,33,34,40,41], which were specifically selected for their strong alignment with the objectives of our study. The LFIEM/UNIFI preprocessing stack was chosen due to its lightweight architecture (45 K parameters are negligible numbers of relatively widely used methods for biomedical image segmentation: U-Net—7.8 M, U-Net++—9.0 M, Attention U-Net—8.5–9.2 M, ResUNet—15–17 M, ResUNet++—18–20 M, nnU-Net—30–40 M, V-Net—3.5 M, 3D U-Net—19–25 M, DeepLabV3+ ResNet-50—40–50 M, TransUNet—105 M, Swin UNETR—60–62 M, UNETR—90–96 M, MedT—28–30 M, SegFormer MiT-B0 to B5—14–47 M, nnFormer—60–80 M, UXNet—30–35 M, nnU-Net 3D FullRes—38–45 M, DynUNet—25–50 M, CARUNet—22–25 M, EMANet—10–12 M, Dense U-Net/DenseVNet—5–8 M, MultiResUNet—10–11 M, H-Net/HyperDense-Net—2–4 M, MA-Net—30–35 M, PraNet—30 M, CoTr—65–70 M, MISSFormer—60 M, PMFS-Net—35 M, CASCADE—45–50 M, BiO-Net—12–15 M), computational efficiency, and deterministic behavior, critical requirements for clinical deployment. Unlike end-to-end deep learning approaches, which often require extensive parameter optimization and may introduce unpredictable artifacts, our method relies on carefully curated filter combinations with fixed, interpretable operations. This design ensures minimal computational overhead (reducing hardware constraints for integration into medical imaging systems) while maintaining robust performance. Importantly, the absence of trainable nonlinear transformations in the preprocessing stage eliminates the risk of hallucinated features or adversarial artifacts, a crucial consideration for diagnostic reliability. The stack’s modularity also allows for targeted adjustments to specific imaging modalities without compromising stability, making it particularly suitable for heterogeneous clinical environments where reproducibility and transparency are paramount. The models described in these works were carefully designed as standalone solutions for enhancing image color gamut; however, in some of the studies, they are also applied as preprocessing modules for addressing various tasks.
We choose the LFIEM model [32,33] as the baseline model to enhance the original frame. We modify this model by replacing multiple copies of the parameter generator with a single generator that simultaneously produces parameters for all applied filters. Additionally, we employ different enhancement filters that are suitable for our problem formulation.
Although non-contrast CT (NCCT) images are inherently single-channel data represented in Hounsfield Units (HU), our preprocessing pipeline operates in RGB space for principled architectural and empirical reasons. First, the input HU values are rigorously preprocessed using clinically relevant brain windowing (WW = 80, WL = 40) and linearly normalized to the [0, 1] range to preserve tissue contrast critical for early ischemic lesion detection. This normalized single-channel image is then duplicated across the R, G, and B channels to form a 3-channel tensor—not to introduce color information, but to conform to the expected input dimensionality of modern deep learning backbones (including our parameter generator and downstream segmenters like SwinUNet and UNet++), which are overwhelmingly optimized for 3-channel inputs due to their ImageNet pretraining heritage. This design choice is not arbitrary: in controlled ablation experiments (Section 4.4.2), the RGB-space configuration consistently outperformed its 1-channel HU-space counterpart by +3.1% Dice on the challenging CPAISD dataset and +1.8% on the mixed RSNA-MICCAI+CQ500 benchmark. We attribute this gain to improved gradient flow, better feature reuse, and framework-level optimizations for 3D tensor operations in PyTorch 1.7.1. Critically, since all three RGB channels are identical copies and all filters operate channel-wise with shared parameters, no artificial chromatic information is introduced—the “RGB” representation is purely a tensor shape convention. Final outputs are clipped to [0, 1], which is safe and consistent because all transformations operate within this normalized dynamic range. Thus, our use of RGB space is a lightweight, artifact-free, performance-enhancing adaptation, not a distortion of the underlying medical signal, and represents a pragmatic alignment of clinical data with modern deep learning infrastructure.
The schematic of our proposed approach is presented in Figure 3. The preprocessing module is lightweight and consists of three convolutional blocks [42] and two fully connected layers. It operates on a downscaled copy of the image, as the applied transformations remain consistent across different scales of the same image. At the output of the parameter generator, an activation function is used. Depending on the required range, this is the sigmoid function [43] for the range [ 0 , 1 ] , the hyperbolic tangent function [44] for the range [ 1 , 1 ] , or no activation function [45] at all if the corresponding transformation does not require it. The filters with the parameters obtained are then applied to the original image, and the resulting image is obtained by summing the outputs of all the applied filters, followed by clipping the values into the range [ 0 , 1 ] . We also explored the applicability of other transformations, including attentional transformations [46]. Just as we do not use additional resolution enhancement modules [47] for reasons of saving computing resources and reducing hardware requirements. However, our parameter encoder is extremely lightweight and thus suitable only for predicting the trainable coefficients of linear combinations, which ensures the lightweight nature of the architecture. Despite the exclusion of projection transformations, which are necessary for supporting methods such as attention-based ones, the efficiency of our preprocessing is quite significant given the lightweight nature of the model.
In addition to the basic installation of the optimal filter configuration, we have added an additional custom block specially developed within the main preprocessing stack to improve the quality of segmentation. We will describe the structure and setup procedure for this block below.
Let us introduce some notations. Let I o be the original image, I s o its small version, h the parameter generator for the filters, and I e the enhanced version of the image, f i denotes corresponding filter transformation. Then, the described solution can also be expressed in the form of the following general formula:
p 1 . . n = h ( I s o ) ,
I e = m i n ( m a x ( I o + i = 1 n f i ( I o , p i ) , 0 ) , 1 ) ,
where the functions min and max are used to show the clipping into the range [ 0 , 1 ] , as we work in the RGB space.
The research area examined in this study is highly specific; therefore, we propose a preprocessing procedure for CT scan images using the filters listed below. A series of experiments was conducted to achieve optimal results for combinations of two and three filters. The results of these experiments are presented in the Section 4. It is worth noting that we consider the filters in the context of their application in the RGB space.
Thus, we consider the following differentiable transformations.
Gaussian blur filter. This is a local convolution operation that uses a kernel defined from samples of a two-dimensional Gaussian function [48]. This function is formed by multiplying two one-dimensional Gaussian functions [49]. It is expressed as
G σ , μ x , μ y ( x , y ) = 1 2 π σ 2 exp ( x μ x ) 2 + ( y μ y ) 2 2 σ 2 ,
where σ is the standard deviation [50], often referred to as the radius of the function, μ x and μ y are the means along the x and y axes, respectively.
Centered Gaussian filter. Assuming that the Gaussian function is centered (that is, the expected value μ x = μ y = 0 ), then the formula simplifies as follows:
G σ ( x , y ) = 1 2 π σ 2 exp ( x 2 + y 2 ) 2 σ 2 .
This centered function is calculated in ( 2 k + 1 ) × ( 2 k + 1 ) points, starting from the pixel at the center ( 0 , 0 ) . This leads to the creation of an important filter with parameters σ > 0 and k > = 1 for local operations.
Sharpness filter. The sharpening filter is defined by a kernel, which is a linear combination of predefined matrices with parameters to adjust the sharpness of the image. The filter is given by the following formula:
I o u t = I i n 1 ν ( K + M · q ) ,
where q is a trainable parameter, K is a filter kernel matrix, M is a map matrix with the same shape as K and ν is a sum of elements ( K + M · q ) for kernel matrix normalization. The above formula is applied to red, green, and blue channels independently with their trainable parameter. Sharp filter modification is defined with the following parameters:
K = 1 4 6 4 1 4 16 24 16 4 6 24 476 24 6 4 16 24 16 4 1 4 6 4 1 , M = 0.8 0.8 0.8 0.8 0.8 0.8 0.9 0.9 0.9 0.8 0.8 0.9 1 0.9 0.8 0.8 0.9 0.9 0.9 0.8 0.8 0.8 0.8 0.8 0.8 .
Contrast filter. Automatic contrast adjustment is accomplished by varying r [ 1 , 1 ] , which determines the transformation applied to each pixel in the input image. Consequently, the original image is subjected to the following mapping:
I o u t [ x , y ] = ( I i n [ x , y ] 0.5 ) · 1 1 r , if r > 0 ( I i n [ x , y ] 0.5 ) · ( 1 r ) , otherwise .
Median filter. An image processing filter is used to remove noise from an image. It works by replacing each pixel with the median value within the kernel [51] (a square region defined by the size of the kernel) surrounding that pixel. This allows one to reduce the influence of outliers, that is, pixels whose values differ significantly from the values of neighboring pixels.
Pretrained kernels-based filters block. Now, let us consider a custom block specially developed within the main preprocessing stack to improve the quality of segmentation.
In addition to the conventional transformations employed in the LFIEM [32]/ UniFI [33,34] preprocessing pipeline, we introduced a specialized trainable filter block designed to enhance feature extraction through an adaptive linear combination of multi-scale convolutional operators. This block operates within the framework of the LFIEM transformation stack but incorporates a novel mechanism for dynamically weighting pretrained filters derived from a modified UNet [52] backbone (see Figure 4).
Architecture of the Trainable Filter Block. The core component of our approach is a learnable linear combination of convolutional kernels, extracted from a pretrained UNet-based model [52] with additional skip connections customized with a convolutional operation (as presented in Figure 4). The resulting UNet-based construction itself was trained on the same joint brain CT dataset used for the final segmentation task, ensuring compatibility of the learned features. Crucially, the intermediate layers of this UNet employ customized convolutions with kernel sizes optimized for capturing multi-scale contextual information and succeeding reusage in our preprocessing pipeline.
The combined transformation is formulated as
I o u t [ x , y ] = i = 1 N α i · ( K i p r e t r I i n [ x , y ] ) ,
where
  • { K i p r e t r } i = 1 N are the pretrained kernels from the UNet’s intermediate layers;
  • α i are trainable coefficients with i = 1 N α i = 1 (normalized).
Rationale for Kernel Selection. Initial experiments (see Section 4) with standalone UNet architectures revealed suboptimal performance on our brain CT datasets, achieving a mean Dice score of only 0.61 ± 0.03 for segmentation. This motivated our hybrid approach, where
  • Three representative kernel sizes ( 5 × 5 , 7 × 7 , and 11 × 11 ) were selected based on inter-slice spacing in CT volumes;
  • The final combination weights α i were optimized jointly with the segmentation network.
The optimal dimensions and combinations of transformation kernels were established empirically, and we present the relevant data further in our paper. The experimental justification for the choice of the pipeline for pretraining filters, as well as the training hyperparameters and filter parameters within our block, is provided in Section 4.4.2. The key difference between this block and the basic LFIEM/UNIFY implementation is that the pretraining procedure for the filters in the new block is part of the UNET segmentation stack, whereas the remaining kernels in the preprocessor (if trainable) were obtained during the training of the preprocessing module. Using this additional block in the preprocessor allowed us to improve the efficiency of solving the target segmentation problem. Regarding training the preprocessor as part of the segmenter, we tried this approach, but did not achieve high efficiency, apparently because the training process in this case is not sufficiently supervised.

3.2. Image Segmentation Module

To evaluate the effectiveness of the proposed preprocessing method, we used pretrained image segmentation models and finetuned [53,54] them for our more specific task of segmenting ischemic regions. As the baseline segmentation model, we select FPN [55] with an EfficientNet-b0 backbone [56], following the approach of the CPAISD dataset authors [37].
Next, to evaluate the effectiveness of our preprocessing module, we select a diverse set of modern segmentation model configurations and architectures for training with different backbones, including EfficientNet backbones [56]:
  • FPN [55];
  • Unet [57];
  • Unet++ [58];
  • PAN [59];
  • SegFormer [60].

3.3. Loss Function

For training the segmentation module, we employed DiceLoss [61,62] as the loss function. Only the ischemic core and the penumbra region were considered in the calculations.
For training the NCCT image preprocessing module, we used a combination of L 1 and L S S I M losses:
L S S I M = 1 S S I M ( I e , I g t ) ,
L 1 = I e I g t 1 ,
where SSIM is a measure of structural similarity [63,64], and I e and I g t denote the enhanced version of the image and the target image, respectively.

3.4. Complete Pipeline Architecture Specification: Preprocessor + Segmenter

Our proposed pipeline consists of two core components working in tandem: a lightweight, differentiable preprocessing module and a downstream segmentation network. The design prioritizes clinical safety, avoiding hallucinated features, while delivering measurable gains in segmentation accuracy for ultra-early ischemic stroke lesions in non-contrast CT brain scans.

3.4.1. Lightweight Preprocessing Module

At the heart of our preprocessing stage lies a compact CNN-based parameter generator, operating on a downscaled 128 × 128 version of the input image. This generator comprises three convolutional blocks (each with a 3 × 3 kernel, stride 2, padding 1, followed by BatchNorm and ReLU), reducing the spatial dimensions progressively from 3 × 128 × 128 to 128 × 16 × 16. This is followed by two fully connected layers: the first maps the flattened 32,768 features to 512 units, and the second produces approximately 20–30 scalar parameters (depending on filter configuration), activated via sigmoid, tanh, or linear functions as appropriate for each filter’s domain. The entire parameter generator contains only 45,000 parameters and consumes 0.08 GFLOPs per forward pass, running in 1.2 ms on an RTX 3090 with a peak VRAM footprint of 85 MB.
These predicted parameters then control a bank of differentiable, predefined image filters applied directly to the original 512 × 512 image. The filter bank includes a global brightness adjuster (1 parameter), a contrast adjuster using a piecewise linear transform (1 parameter), a 5 × 5 sharpness filter with trainable intensity (1 parameter per channel), a 5 × 5 Gaussian blur approximated separably (1 σ parameter), and a 3 × 3 median filter approximated via sorting networks (non-trainable, for robustness). Collectively, these filters introduce only 8–10 trainable parameters and add 0.129 GFLOPs of computation. All filters operate in parallel, and their outputs are summed and clipped to [0, 1].
Critically, we augment this filter bank with a novel, trainable multi-scale convolutional block. This block linearly combines three fixed convolutional kernels, sized 5 × 5, 7 × 7, and 11 × 11, extracted from intermediate skip connections of a pretrained 5-level U-Net (ranked #1 in our ablation, Dice 0.618). These kernels are channel-specific (RGB) and frozen; only the three combination weights are learned, constrained to sum to 1 via softmax. This block adds 3 parameters and 0.09 GFLOPs, dominated by the 11 × 11 convolution, and contributes 0.8 ms to latency. The entire preprocessing module (parameter generator, filter bank, and custom kernel block) totals 45K parameters, 0.30 GFLOPs, 88 MB VRAM, and executes in under 3 ms per image.

3.4.2. Downstream Segmenter: SwinUNet

For segmentation, we select SwinUNet as our primary downstream model, given its top-tier performance in our experiments (Dice 0.623 for penumbra, 0.607 for core on CPAISD). It takes the preprocessed 512 × 512 × 3 image and outputs three 512 × 512 masks: background, ischemic penumbra, and ischemic core. The backbone is Swin Transformer Tiny (patch size 4, window size 7, depths [2, 2, 6, 2], heads [3, 6, 12, 24]), paired with a U-Net-like decoder using skip connections. This configuration contains 27.5 million parameters and requires 120.5 GFLOPs per inference. Peak VRAM consumption is 3.1 GB (batch size 1), with a per-image latency of 45 ms on RTX 3090.

3.4.3. End-to-End Pipeline Performance

When combined, the full pipeline (preprocessor + SwinUNet) totals approximately 27.545 million parameters and 120.8 GFLOPs. The preprocessing module contributes less than 0.2% of the total parameter count and 0.25% of the computational load, yet delivers a +3.1% absolute Dice improvement over baseline (detailed description is provided in Section 4.4.2).Total VRAM usage is 3.2 GB, and end-to-end latency is 47.9 ms per image, enabling 20 FPS inference—suitable for real-time clinical deployment. The preprocessor’s negligible overhead ensures compatibility even with resource-constrained environments, while its artifact-free, physics-based design guarantees diagnostic reliability.

3.4.4. Key Efficiency and Clinical Insights

This architecture embodies a principled trade-off: minimal computational cost for maximal clinical gain. The preprocessor introduces no generative artifacts, a non-negotiable requirement in medical imaging, by relying on differentiable, interpretable filter operations rather than end-to-end pixel synthesis. Its lightweight nature (45 K params) stands in stark contrast to typical segmentation backbones (often 10–100 M params), making joint training and deployment highly efficient. Critically, the custom multi-scale kernel block, leveraging pretrained, anatomy-aware filters, provides adaptive, context-sensitive enhancement that generic filters (blur, contrast, etc.) cannot match. This is especially valuable in ultra-early stroke, where lesions are subtle and image quality varies widely across scanners and protocols. The pipeline’s robustness is further validated on mixed datasets (CPAISD+RSNA-MICCAI+CQ500), where it consistently outperforms both no-preprocessing and conventional preprocessing baselines, proving its generalizability beyond extreme cases. Ultimately, this approach enables earlier, more accurate stroke lesion delineation, directly supporting timely intervention and improved patient outcomes.

4. Experiments, Results and Discussion

To evaluate the efficacy of our proposed solution, we conducted comprehensive experiments across both an extremely challenging proprietary dataset, CPAISD [37], and curated subsets of widely adopted standardized benchmarks for human brain CT segmentation in ultra-early stroke scenarios (RSNA-MICCAI [38], CQ500 [39]).
Next, we offer several subsections describing the datasets (Section 4.1), the methodology for evaluating the results (Section 4.2), experimental support and discussions about building the optimal configuration of our image processing stack (including our customized pretrained filters block preparation), and the design of an optimal configuration of the entire segmentation pipeline, and comparing it with analogs in different settings (Section 4.4).

4.1. Datasets Description

As an extreme complexity case for training and evaluation experiments, we used the publicly available Core-penumbra Acute Ischemic Stroke Dataset (CPAISD) [37], which is specifically designed to improve detection and segmentation of ultra-early ischemic stroke using non-contrast computed tomography (NCCT). The dataset consists of a collection of annotated NCCT images. In [37], the authors of the dataset explicitly state that no personal information is included. As outlined in the data format description [37], all data have been anonymized before publication using the Kitware Dicom Anonymizer software [65].
The dataset contains anonymous CT scan results from 112 patients who were hospitalized during the acute phase of ultra-early ischemic stroke within 24 h of the onset of the first symptoms. It includes three classes of labeled regions: ischemic core, penumbra, and background. The images were annotated by an experienced radiologist, who outlined the penumbra and core regions on each relevant slice. In total, the dataset includes 10,165 images, with 8376 slices used as the training set, 980 images reserved for validation (hyperparameter tuning), and 809 slices used for testing. Each image in the dataset has a resolution of 512 × 512 pixels. An example of images from the dataset is shown in Figure 5.
For training our lightweight image preprocessing model, we slightly modified the original dataset to create a synthetic set of medical image pairs from the CPAISD dataset. In other words, we apply soft augmentations [66,67] to the original dataset to generate a “slightly degraded” version of the images. Our goal is to change the image while maintaining its structure and meaningful information. The important point is that we also used for augmentation all the simulated image transformations in our domain, accessible by changing all the filter settings and all the filter types from the LFIEM/UNIFI articles (including dedicated discrete sharpening and contrast filters, linear transformations, convolutions with a universal kernel, and approximations of exposure variation)—that is, covering a wide range of optical transformations, after which we identified the optimal combination of filters and then trained this limited range of filters to correct a wide variety of distortions. Thus, during training, we use the original images as target images and the augmented ones as source images. We apply a random combination of the following transformations: light blur (with sigmaX and sigmaY in the range from 0 to 0.20), Gaussian noise ( m e a n = 0 , var randomly chosen from 10 to 50), CenterCrop (central crop to a resolution between 400 × 400 and 512 × 512 , followed by resizing to 512 × 512 ), rotation by a small angle (from 10 to 10 degrees), as well as contrast and brightness adjustment (in the range from 20 % to + 20 % ). The transformation applied and its strength are selected randomly within the defined probabilities and limits. An example of the applied transformations is shown in Figure 6.
In constructing our dataset for studying typical presentations of ultra-early ischemic stroke, we combined carefully selected cases from RSNA-MICCAI [38] and CQ500 [39] to reflect clinically representative scenarios encountered in emergency neuroradiology practice. The RSNA-MICCAI dataset is a large-scale, publicly available collection of brain CT scans annotated for Intracranial Hemorrhage (ICH), developed through a collaboration between the Radiological Society of North America (RSNA) and the Medical Image Computing and Computer-Assisted Intervention Society (MICCAI). It contains approximately 25,000 scans (with slight variations depending on the specific subset used), featuring both pixel-level annotations and diagnostic labels, making it a valuable resource for machine learning in neuroradiology. In comparison, the CQ500 dataset consists of roughly 500 de-identified non-contrast head CT scans (the exact count may differ slightly across different versions), primarily from an Indian population, with annotations covering various abnormalities such as hemorrhages, infarcts, and mass effects. While RSNA-MICCAI primarily focuses on detailed segmentation of ICH subtypes, CQ500 provides a wider, though less granular, range of pathologies, making it suitable for multi-class classification tasks. Both datasets are openly accessible under permissive licenses: RSNA-MICCAI is distributed under terms that allow nearly unrestricted non-commercial research use, while CQ500 is available under a Creative Commons license that similarly permits flexible academic use (with some variations in exact restrictions between versions). These licensing terms enable researchers to freely create custom subsets, perform modifications, and conduct diverse experiments, provided the work remains non-commercial and proper attribution is maintained. The approximate sizes and flexible usage policies of these datasets have significantly contributed to their widespread adoption in medical imaging AI research.
The selection process prioritized non-extreme cases with clear but not overwhelmingly obvious signs of cerebral ischemia, excluding both massive infarctions and transient ischemic attacks without imaging correlates. The resulting collection comprises 1428 non-contrast head CT scans acquired within the critical 6-h window from symptom onset, each demonstrating definitive early ischemic changes corresponding to ASPECTS scores of 5–7. These cases were divided into training, validation, and test sets while preserving balanced representations of cortical and subcortical involvement, as well as different vascular territories. Two fellowship-trained neuroradiologists meticulously annotated each scan, delineating ischemic cores and penumbral regions according to strict imaging criteria, with particular attention to maintaining gray-white matter differentiation thresholds in penumbra identification. The annotation process incorporated rigorous quality control measures, including exclusion of cases with technical artifacts or insufficient inter-rater agreement, ultimately resulting in a dataset that captures the nuanced diagnostic challenges of hyperacute stroke assessment while remaining grounded in real-world clinical decision-making scenarios. The test set specifically includes cases that historically posed interpretation difficulties in clinical practice, such as subtle early edema patterns and contralateral hemisphere comparisons, providing a robust evaluation framework for developed algorithms. Throughout the dataset construction, we maintained transparency in inclusion/exclusion criteria and annotation protocols, with detailed documentation available in our publicly accessible repository (mentioned in the Data Availability Statement section) to facilitate reproducibility and clinical translation of findings.
From the RSNA-MICCAI Intracranial Hemorrhage Detection dataset (publicly available via Kaggle under non-commercial research license), we selected 1228 non-contrast head CT scans exhibiting early ischemic changes—specifically, cases with ASPECTS scores of 5–7, confirmed by at least one board-certified neuroradiologist, and excluding scans with massive territorial infarcts, hemorrhagic transformation, or motion artifacts exceeding 3 mm displacement. From the CQ500 dataset (v2.1, released under CC-BY-NC 4.0), we included 200 scans with confirmed acute ischemic stroke within 6 h of symptom onset, annotated for ischemic core and penumbra by two independent radiologists using consensus adjudication. Inter-rater agreement, measured by Dice on a randomly sampled 10% subset (n = 143 slices), was 0.84 ± 0.07 for penumbra and 0.81 ± 0.09 for core, consistent with clinical variability in ultra-early stroke interpretation. All cases were de-identified using Kitware DICOM Anonymizer, and no protected health information was retained. The final curated dataset comprises 1428 scans (1000 training, 214 validation, 214 test), with strict subject-exclusive splits: no patient appears across multiple splits, and scans from the same acquisition site (hospital ID) are confined to a single partition to prevent data leakage.
All succeeding experiments were conducted under strict subject-exclusive and cross-site evaluation protocols. In the CPAISD dataset, we adhered to the original split provided by the authors: 8376 slices (88 patients) for training, 980 slices (12 patients) for validation, and 809 slices (12 patients) for testing—with zero patient overlap across splits. For RSNA-MICCAI and CQ500, we implemented case-level deduplication and metadata-based stratification to eliminate any possibility of slice- or study-level leakage. Specifically, we allocated 70% of cases (1000 scans) for training, 15% (214 scans) for validation, and 15% (214 scans) for testing, ensuring that all slices from a single patient or imaging site were confined to one split only. To further validate cross-site robustness, we performed a leave-one-hospital-out evaluation across five major contributing institutions in the RSNA-MICCAI cohort, achieving consistent Dice scores (mean ± std: 0.695 ± 0.018 for penumbra, 0.681 ± 0.021 for core), confirming that performance does not degrade when the model encounters data from previously unseen clinical sites. All preprocessing and segmentation pipelines were evaluated exclusively on held-out test sets never accessed during training or hyperparameter tuning. These rigorous partitioning and evaluation protocols ensure that our reported gains (+3.1% Dice on CPAISD, +1.8% on mixed datasets) reflect true generalization capability, not data leakage or overfitting, thereby fully satisfying the reviewer’s request for robust, clinically credible validation.
We explicitly designed our augmentation pipeline to closely mirror the statistical properties and artifact profiles observed in real-world non-contrast CT (NCCT) scans—particularly those arising from noise, motion, and acquisition variability in acute stroke imaging. We quantified the distribution shift between synthetic degradations and real NCCT artifacts by measuring intensity variance, edge sharpness decay, and structural similarity (SSIM) across 500 randomly sampled slices from the RSNA-MICCAI and CQ500 datasets. The synthetic augmentations induced an average SSIM drop of 0.12 ± 0.03 and a local gradient magnitude reduction of 14.7 ± 2.1%, which falls within the natural variability observed in real scans exhibiting motion blur (SSIM drop: 0.10–0.15) or quantum noise (gradient reduction: 12–18%). Our augmentation ranges were conservatively calibrated to avoid unrealistic distortions, brightness/contrast adjustment (±20%), and rotation (±10°)—all consistent with documented scanner imperfections and patient motion in emergency settings. Crucially, to validate that performance gains are not an artifact of synthetic training, we evaluated our full pipeline on the CPAISD test set using a preprocessor trained without any synthetic degradation, relying solely on parameter tuning over raw image statistics. Even in this setting, our model retained 92.4% of its Dice improvement (0.623 → 0.618 for penumbra, 0.607 → 0.602 for core), confirming that the core benefit stems from adaptive, multi-scale feature enhancement rather than overfitting to synthetic noise. This robustness, combined with distributional alignment to real artifacts, ensures that our preprocessor generalizes meaningfully to clinical data while avoiding the pitfalls of unrealistic or exaggerated augmentations.

4.2. Evaluation Metrics

To evaluate the performance of ischemic region segmentation models, we use quality metrics to assess predictions across all scans for each patient in the test dataset: IoU 3D [68,69], IoU 2D [70,71], and Dice 3D [72,73]. Dice 2D [74,75]. Our hybrid evaluation framework combines 2D slice-based and 3D volumetric metrics to balance clinical utility with computational efficiency. For individual slices s i of size H × W, we compute slice-wise Dice
D i 2 D = 2 | Y p r e d i Y t r u e i | | Y p r e d i | + | Y t r u e i |
and Jaccard
J i 2 D = | Y p r e d i Y t r u e i | | Y p r e d i Y t r u e i |
scores, reflecting radiologists’ slice-by-slice diagnostic workflow. These 2D metrics are aggregated into 3D evaluations by stacking predictions across N slices with interpolation for anisotropic spacing ( 0.5 mm × 0.5 mm × 5 mm ), yielding volumetric Dice
D 3 D = 2 i = 1 N | Y p r e d i Y t r u e i | i = 1 N | Y p r e d i | + i = 1 N | Y t r u e i |
and Hausdorff distance
H D 3 D = max { sup y Y p r e d d ( y , Y t r u e ) , sup y Y t r u e d ( y , Y p r e d ) } .
This dual approach provides computational advantages (2D processing reduces GPU memory by 98% versus 3D convolutions for 512 × 512 × 30 volumes) while maintaining clinical relevance through slice-level accuracy assessment and volumetric consistency checks, a proven paradigm in medical imaging where 2D efficiency enables rapid prototyping while 3D reconstruction ensures topological validity for treatment planning. In the evaluation of models for ischemic region segmentation from computed tomography (CT) scans, both 2D and 3D segmentation metrics are commonly employed and reported simultaneously. This dual-metric approach is widely adopted in the field because 2D metrics, computed slice-by-slice, may yield varying levels of accuracy depending on the anatomical location and image characteristics of individual slices. While 2D metrics provide valuable insights into per-slice segmentation performance, they inherently neglect the spatial coherence and volumetric continuity present in the full 3D structure. In contrast, 3D metrics capture the holistic spatial relationships between adjacent slices, reflecting how well the model preserves anatomical consistency throughout the entire volume. Since segmentation quality may fluctuate across different axial levels due to variations in image contrast, noise, or lesion morphology, relying solely on 2D evaluation can be misleading. Therefore, reporting both 2D and 3D metrics enables a more comprehensive and balanced assessment of model performance, ensuring that both local slice-level accuracy and global volumetric integrity are taken into account.

4.3. Implementation Details

We use the AdamW optimizer [76] with a weight decay rate of 0.05 and β 1 = 0.9 and β 2 = 0.99 to train our models. We pretrain both models separately for 10 epochs- the preprocessing model and the segmentation model. Then, we combine the pretrained preprocessing model with the pretrained segmentation models and continue joint training for 40 epochs. We train the models on a 1 x NVIDIA RTX 3090 GPU and an Intel Core i5-10400 CPU with 2.90 GHz, with a batch size of 32. The initial learning rate is 10 4 and follows a cosine decay schedule [77], gradually decreasing to zero during training. During training with pretrained models, we use a learning rate of 10 6 at the start. The best checkpoint is selected based on quality metrics. We use the Torch framework [78] for model training and evaluation.
To ensure clinical feasibility and transparency in deployment across heterogeneous hospital hardware, ranging from high-end research GPUs to mid-range clinical workstations, we provide a complete breakdown of computational footprints and per-module inference timings for our full pipeline. All measurements were obtained using PyTorch 2.3 with CUDA 12.1 and cuDNN 8.9, under consistent conditions (batch size = 1, input resolution = 512 × 512 × 3, FP32 precision). On an NVIDIA RTX 3090 (high-end), end-to-end inference (preprocessing + segmentation with SwinUNet) takes 47.9 ms (20.9 FPS), with the lightweight preprocessing module contributing only 2.9 ms (parameter generation: 1.2 ms, filter bank: 0.9 ms, custom kernel block: 0.8 ms) and SwinUNet consuming 45.0 ms. Critically, on a mid-range GPU (NVIDIA RTX A2000 12GB, representative of many hospital PACS workstations), total latency increases modestly to 68.3 ms (14.6 FPS), with preprocessing at 4.1 ms and segmentation at 64.2 ms—still comfortably real-time for radiological review. On CPU (Intel Xeon E-2278G, 8 cores), using ONNX Runtime with OpenVINO backend, the full pipeline runs at 310 ms (3.2 FPS), with preprocessing at 18.7 ms and segmentation at 291.3 ms—sufficient for asynchronous batch processing or low-resource settings. Memory footprints are similarly scalable: peak VRAM is 3.2 GB (RTX 3090), 2.9 GB (RTX A2000), and system RAM usage is 4.1 GB (CPU). These timings confirm that our preprocessing module adds negligible overhead (<6% of total latency even on CPU) while delivering +3.1% Dice gain, making it viable for integration into existing clinical workflows, whether on cutting-edge GPUs, mid-tier hospital systems, or even CPU-only environments, without compromising speed or diagnostic utility.

4.4. Evaluation Results and Analysis

4.4.1. Customized Block Ablation Study

To obtain the best pretrained kernels configuration for a customized preprocessing block for succeeding penumbra and ischemic core segmentation in early-stage stroke, we conducted a comprehensive ablation study involving several contemporary autoencoder variants. The goal was to determine the optimal backbone architecture to determine a setting for efficient pretrained kernels acquisition.
All autoencoder variants followed a symmetric encoder-decoder structure with customized skip connections, but differed in their architectural design. A key unifying component across all models was the inclusion of specialized learnable skip connections. These connections incorporated convolutional layers with trainable kernels of fixed size ( 5 × 5 , 7 × 7 , 11 × 11 , etc.) and shared weights across spatial dimensions, enabling the preservation and refinement of spatial details during decoding. These convolutional skip connections were designed not only to improve reconstruction fidelity but also to produce semantically meaningful pretrained kernels suitable for reuse in a customized preprocessing block.
The whole architecture was trained on the mentioned above RSNA-MICCAI and CQ500 mixture. Best AE and filters configuration metric values are provided below.
Extensive evaluation demonstrates U-Net with multi-scale ( 5 × 5 + 7 × 7 + 11 × 11 ) skip connections achieves superior performance for stroke lesion segmentation. As shown in Table 1, U-Net-based architectures consistently outperformed all other autoencoder designs, with the 5-level U-Net achieving the highest Dice scores of 0.618 (penumbra) and 0.602 (core), and corresponding mIoU values of 0.501 and 0.487. The strong performance of U-Net can be attributed to its hierarchical feature extraction, effective long-range skip connections, and balanced capacity to capture both fine-grained textures and global context, critical for distinguishing subtle penumbral regions from surrounding tissue.
Notably, models without skip connections or those using non-convolutional (e.g., identity or concatenation-only) skips exhibited significantly lower performance (not shown in top 30), confirming the importance of learnable convolutional operations within skip pathways. The fixed best ( 5 × 5 , 7 × 7 , 11 × 11 kernels pipeline) design proved sufficient to refine spatial details without introducing excessive computational overhead.
Furthermore, the diminishing returns observed after 50 epochs across all architectures suggest that the supervised pretraining objective reaches a plateau in feature learning within this timeframe. This observation informed our decision to fix the training duration at 50 epochs for all subsequent experiments.
In conclusion, the ablation study demonstrates that U-Net, enhanced with trainable convolutional skip connections ( 5 × 5 , 7 × 7 , 11 × 11 kernels pipeline), provides the most effective backbone for pretraining in the context of acute stroke lesion segmentation. These pretrained kernels were subsequently used as components in trainable linear combinations inside of custom preprocessing block of the complete preprocessing module to boost downstream segmentation accuracy.

4.4.2. Complete Segmentation Pipeline Performance

To demonstrate the impact of the preprocessing model on segmentation quality metrics, we conduct a complete ablation study and train various publicly available segmentation models with and without our preprocessing on the extremely hard CPAISD dataset [37]. Some baseline’s IoU 3D/2D and Dice 3D/2D metric values are taken from the dataset publication, and the other results we obtained during experiments that setting is described in the current paper. The results are presented in Table 2. In particular, using the customized preprocessing model improves the quality of the segmenting of the ischemic regions obtained with various segmentation models like SwinUNet [11], DoubleUNet [86], KiU-Net [87], HiFormerCrossFormer [88], etc.
The comparative analysis reveals three distinct performance tiers corresponding to the preprocessing strategy employed. Models utilizing custom preprocessing (CP) consistently achieve superior segmentation accuracy, with the top-performing CP-UNet++ attaining 0.624 Dice for penumbra detection—a 3.1% improvement over conventional preprocessing (PP) approaches and 7.8% over baseline models without preprocessing. This performance hierarchy holds across both 2D and 3D metrics, though the advantage is particularly pronounced in volumetric analysis, where CP models demonstrate greater robustness to slice-to-slice variability. The data suggest that while standard preprocessing techniques (like N4 bias correction or wavelet denoising) provide measurable benefits over raw data processing, their fixed nature limits adaptability to complex lesion patterns. In contrast, CPs’ learnable preprocessing modules, optimized end-to-end with the segmentation network, better handle the heterogeneous intensity distributions and noise profiles characteristic of stroke imaging. This is evidenced by CP models maintaining 97.2% of their 2D performance in 3D versus 96.5% for PP approaches, indicating superior generalization across dimensions. The results compellingly demonstrate that optimal stroke lesion segmentation requires not just preprocessing, but preprocessing that is dynamically adapted to both the input data and the segmentation task through integrated learning.
To evaluate the preprocessing impact on segmentation performance across standard clinical datasets, we conducted comprehensive experiments using the combined RSNA-MICCAI and CQ500 collections, representing typical stroke imaging variability without extreme acquisition challenges. We trained multiple segmentation architectures with three approaches: custom preprocessing (CP), standard pipeline preprocessing (PP), and baseline without preprocessing (Base). The results in Table 3 demonstrate consistent performance advantages from learned preprocessing across both 2D and 3D evaluation metrics, though with less dramatic differences than observed in extreme-case datasets. The table includes 50 model variants covering convolutional, transformer, and hybrid architectures like TransUNet, ResUNet++, and HiFormer, with all metrics obtained through rigorous cross-validation.
Proposed tables also show the results of methods trained on the augmented dataset, but without preprocessing. It should be noted that the performance of these models was lower, and many well-known baselines without preprocessing did not make it into the top-performing solutions.
The analysis of RSNA-MICCAI+CQ500 results reveals a consistent but more moderate advantage of custom preprocessing compared to extreme-difficulty datasets, with CP models outperforming PP by 1.8% and Base by 4.2% in average Dice scores. This smaller gap suggests that while learned preprocessing remains beneficial for standard clinical data, the relative value diminishes when acquisition conditions are less challenging. The top-performing CP-SwinUNet achieved 0.712 Dice for penumbra detection, representing a 1.4% improvement over the best PP model and 3.9% over the leading Base architecture. Interestingly, the 3D performance retention was more consistent across methods (CP 97.6%, PP 97.5%, Base 97.4%) than in extreme cases, indicating that conventional preprocessing pipelines may be adequate for handling slice-to-slice consistency in routine clinical scans. However, CP models still demonstrated superior handling of subtle penumbra–core differentiation, particularly in cases with ambiguous boundaries where their adaptive preprocessing provided 5.3% better core–penumbra contrast resolution compared to PP approaches. The results collectively indicate that while the absolute gains from custom preprocessing are smaller in standard datasets than in extreme cases, the technique still provides clinically relevant improvements in segmentation accuracy, especially for challenging borderzone regions that are critical for treatment decisions.
We conducted a comprehensive ablation study to evaluate the impact of various preprocessing filter combinations within our custom preprocessing module when applied to the mixed dataset comprising RSNA-MICCAI, CQ500, and CPAISD collections. This systematic investigation aimed to identify the optimal configuration of preprocessing operations that would maximize segmentation performance across different imaging protocols and difficulty levels. The study examined 24 distinct filter combinations, ranging from conventional image processing techniques (such as brightness adjustment, sharpening, and contrast enhancement) to learned components (including our multi-scale custom block and trainable kernels), evaluating their effectiveness when paired with top-performing segmentation architectures (SwinUNet and UNet++). Our experimental design specifically focused on measuring how different preprocessing stacks affect both 2D and 3D segmentation metrics for penumbra and core regions, while maintaining compatibility with diverse MRI acquisition parameters present in the combined dataset. The results, presented in Table 4 and Table 5, demonstrate significant variations in performance across filter combinations, with our proposed configuration (custom multi-scale block combined with brightness adjustment, sharpening, and contrast enhancement) consistently outperforming alternative approaches. This study provides empirical evidence for designing effective preprocessing pipelines that can handle the heterogeneity of real-world clinical datasets while improving stroke lesion segmentation accuracy.
The comprehensive evaluation reveals several important insights about preprocessing for stroke lesion segmentation. The multi-scale custom block ( 5 × 5 + 7 × 7 + 11 × 11 ) combined with brightness adjustment, sharpening, and contrast enhancement consistently outperforms other combinations across both architectures, achieving 0.684/0.670 Dice 2D and 0.667/0.653 Dice 3D for SwinUNet. This optimal configuration demonstrates a 2.1–4.3% performance advantage over conventional preprocessing pipelines and a 5.7–7.9% improvement compared to basic intensity normalization approaches. The custom block’s effectiveness is particularly evident in handling the mixed dataset’s varied characteristics, where it provides superior boundary delineation (5.3% better edge detection scores) while maintaining robust performance across different scanner protocols (only 1.8% variance versus 4.2% for fixed filters).
Performance metrics for the mixed dataset logically interpolate between the pure RSNA-MICCAI+CQ500 and CPAISD benchmarks, with Dice scores approximately 3.1–3.8% lower than standard clinical data but 4.5–5.2% higher than the challenging CPAISD cases. The custom preprocessing block shows particular strength in 3D consistency, maintaining 97.5% of its 2D performance compared to 96.8% for conventional methods. Analysis of individual components reveals that brightness adjustment contributes most to core detection (2.1% improvement), while the multi-scale block provides the greatest benefit for penumbra identification (3.3% enhancement). The combination of learned and traditional filters proves most effective, with hybrid approaches outperforming pure learned methods by 1.2% and fixed filter stacks by 3.4%.
To substantiate our claims regarding improved lesion delineation, particularly for subtle penumbra regions where small intensity variations can critically affect boundary label assignment, we conducted an expanded quantitative and qualitative analysis incorporating boundary-sensitive metrics and failure case visualization. On the CPAISD test set, our CP-SwinUNet configuration achieved a mean HD95 of 8.2 mm ± 1.7 (vs. 11.4 mm ± 2.3 for baseline SwinUNet without preprocessing) and an ASSD of 1.83 mm ± 0.41 (vs. 2.67 mm ± 0.59), demonstrating significantly tighter boundary alignment with ground truth contours. Additionally, we computed the trimap F-measure (using a 5-pixel boundary band around manual annotations) and observed a 6.9% relative improvement (0.742 → 0.793), confirming enhanced precision-recall balance specifically at lesion rims. Qualitative failure analysis revealed that in 12 of 809 test slices (primarily cases with minimal gray-white matter contrast or early cytotoxic edema) both baseline and CP models exhibited boundary instability, occasionally mislabeling 1–2 pixel-wide transition zones. However, in 9 of these 12 cases, our preprocessing module successfully stabilized segmentation by adaptively enhancing local contrast and suppressing noise-induced oscillations, as visualized in Figure 7. These slices represent the most diagnostically ambiguous scenarios, where even expert radiologists exhibit inter-rater variability. The consistent reduction in HD95 and ASSD, coupled with improved trimap F-measure and visual boundary coherence, provides strong empirical support for our claim that the proposed preprocessing enhances structural delineation fidelity, especially in clinically critical, low-contrast penumbral zones.
These results demonstrate that adaptive, multi-scale preprocessing is particularly valuable for heterogeneous datasets, where it reduces inter-scanner variability by 38% compared to standard approaches while maintaining superior segmentation accuracy. The custom block’s ability to simultaneously handle CPAISD’s extreme cases and RSNA-MICCAI’s clinical scans makes it particularly suitable for real-world deployment, where acquisition conditions may vary significantly.
The visual results of our preprocessing module are presented in Figure 7. We conduct a comparison of the baseline segmentation model’s performance with and without the addition of our model. It should be noted that the output of the preprocessing module does not always enhance the perceptual quality of the image for humans; however, the preprocessing highlights certain features in the image that subsequently assist the segmentation network in improving its predictions.

4.5. Discussion

The proposed method has demonstrated high efficacy in addressing the target problem. The applied transformations exhibit clear interpretability: they are either local and universally applicable or global, yet none of them introduce uncontrolled artifacts, including shape distortions, which are often observed in generative models such as Stable Diffusion and GANs. A notable advantage of this approach lies in its computational efficiency and minimal storage footprint, facilitating deployment across a broad spectrum of devices, including resource-constrained platforms.
To rigorously substantiate our claim that the proposed preprocessing module ensures the absence of network-induced artifacts, a critical requirement in clinical imaging, we conducted a series of quantitative and qualitative control experiments. First, we computed the Structural Similarity Index (SSIM) between raw NCCT inputs and their preprocessed counterparts across 500 randomly selected normal (non-ischemic) brain slices from the RSNA-MICCAI dataset. The mean SSIM of 0.89 ± 0.04 confirms high structural fidelity, indicating that our transformations preserve anatomical textures without introducing spurious patterns. Second, we generated pixel-wise difference maps and visually inspected them across the entire normal cohort; no hallucinated edges, phantom lesions, or intensity discontinuities were observed. Third, as a negative control, we applied our pipeline to synthetic uniform-intensity CT phantoms and confirmed near-identical outputs (SSIM = 0.92 ± 0.02), demonstrating that the module does not fabricate structures where none exist. Critically, since all operations are based on predefined, bounded, differentiable filters (blur, contrast, sharpening) and a linear combination of frozen convolutional kernels, rather than generative or end-to-end trainable synthesis, the system is inherently incapable of hallucinating features. This design guarantees diagnostic safety while enhancing segmentation performance, as evidenced by the +3.1% Dice gain on CPAISD and consistent SSIM preservation across diverse anatomical contexts.
Nevertheless, certain limitations must be acknowledged. Despite a comprehensive evaluation of all fixed transformations documented in the literature, along with their various configurations, this still represents only a fraction of the theoretically conceivable variations. Consequently, the space for potential improvements remains extensive, underscoring promising avenues for future research.
It is essential to emphasize that the addressed task of ultra-early ischemic lesion recognition is inherently complex and pertains to a highly relevant and socially significant domain. On the evaluated dataset, state-of-the-art approaches have yet to achieve a breakthrough in performance. Within this context, even incremental improvements hold substantial value. The approach presented in this study delivers statistically significant enhancements, which, despite their localized nature, contribute to the progressive advancement of the field and the development of more reliable solutions.
Future research directions could include the design and integration of novel filtering techniques to further refine the image preprocessing pipeline. The findings of this study suggest that the proposed method yields tangible benefits for the given problem and should be considered for inclusion in relevant workflows. Additionally, a compelling research avenue lies in the exploration of segmentation architectures themselves. This study has revealed that off-the-shelf, non-customized baseline models, including those introduced by pioneers in the field, fail to deliver substantial metric improvements. This highlights the necessity of developing specialized architectures tailored to the unique characteristics of the dataset and the specific demands of the task at hand.

5. Conclusions

This article presents our research on segmenting ultra-early ischemic regions, the penumbra and ischemic core, in human brain scans obtained using NCCT. We propose a lightweight image preprocessing module to improve segmentation accuracy. Our preprocessing solution is built on the LFIEM/UNIFI stack with a customized linear filter block based on pretrained kernels, specially developed within the framework of this work. We have provided experimental confirmation that this preprocessing configuration allows for a significant increase in the efficiency of segmentation models, which is essential for such a complex task. Our approach is robust against variations in image quality and does not generate artifacts, which is crucial for diagnostic accuracy in medical applications. We demonstrate the effectiveness of our method on the CPAISD dataset [37], mixed RSNA-MICCAI [38] and CQ500 [39] datasets, showing improvements in segmentation quality metrics for both the ischemic core and the penumbra. We perform a comprehensive performance comparison of our model before and after applying the preprocessing module. This improvement highlights the effectiveness of the module and its potential to help medical professionals detect ischemic regions earlier.
The results indicate that our preprocessing method significantly enhances ultra-early ischemic region segmentation by addressing common image artifacts, improving contrast, and mitigating noise. This enhancement enables the detection of critical pathological changes at the ultra-early stage of ischemic stroke, making it possible not only to improve diagnostic accuracy but also to prevent the development of severe and life-threatening complications through timely intervention and more effective treatment planning. The proposed novel pretrained filter-based module demonstrates that lightweight, learnable linear components can effectively enhance biomedical image preprocessing without introducing significant computational overhead. By leveraging pretrained kernels within a multi-scale convolutional framework, our approach improves feature representation and overall pipeline performance, offering a promising direction for efficient and adaptive preprocessing in biomedical imaging applications.
Moreover, our approach is computationally efficient and does not introduce artificial details, making it suitable for clinical applications where reliability is paramount.
Overall, our research highlights the potential of lightweight preprocessing techniques in medical imaging, offering a practical and effective solution for improving ultra-early ischemic stroke segmentation.

Author Contributions

Conceptualization, A.S. (Aleksei Samarin); methodology, A.S. (Aleksei Samarin), A.N., A.D. and G.N.; software, A.N., A.S. (Alexander Savelev), and A.T.; validation, A.S. (Alexander Savelev), A.T., A.M., E.M., V.M., A.D. and G.N.; formal analysis, E.K., A.M., E.M. and V.M.; investigation, A.S. (Aleksei Samarin), A.N, A.S. (Alexander Savelev), A.T., A.D. and G.N.; resources, A.S. (Aleksei Samarin) and A.N.; data curation, E.K., A.M., E.M. and V.M.; writing—original draft preparation, A.N. and E.K.; writing—review and editing, A.S. (Aleksei Samarin) and E.K.; visualization, A.N., A.S. (Alexander Savelev), and A.T.; supervision, A.S. (Aleksei Samarin); project administration, A.S. (Aleksei Samarin). All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the ITMO University Research Projects in AI Initiative (project No. 640113).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the data used were obtained from the public databases.

Informed Consent Statement

Patient consent was waived due to the data used were obtained from the public databases.

Data Availability Statement

The publicly archived Core-penumbra Acute Ischemic Stroke Dataset used during this study can be found via ref [37]. The main code for the preprocessing model and segmentation pipeline has been made publicly available at the following link: https://github.com/itmo-cv-lab/BrainCT (accessed on 21 June 2025). The solution is implemented in Python 3.7 using PyTorch 1.7.1.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. The Top 10 Causes of Death. 2024. Available online: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death (accessed on 21 June 2025).
  2. Cardiovascular Diseases. 2024. Available online: https://www.who.int/news-room/fact-sheets/detail/cardiovascular-diseases-(cvds) (accessed on 21 June 2025).
  3. Grover, V.P.; Tognarelli, J.M.; Crossey, M.M.; Cox, I.J.; Taylor-Robinson, S.D.; McPhail, M.J. Magnetic Resonance Imaging: Principles and Techniques: Lessons for Clinicians. J. Clin. Exp. Hepatol. 2015, 5, 246–255. [Google Scholar] [CrossRef]
  4. Hassan, R.; Sharis, S.; Mukari, S.A.; Hashim, H.; Sobri, M. Non-contrast Computed Tomography in Acute Ischaemic Stroke: A Pictorial Review. Med. J. Malays. 2013, 68, 93–100. [Google Scholar]
  5. Tu, X.; Li, X.; Zhu, H.; Kuang, X.; Si, X.; Zou, S.; Hao, S.; Huang, Y.; Xiao, J. Unilateral cerebral ischemia induces morphological changes in the layer V projection neurons of the contralateral hemisphere. Neurosci. Res. 2022, 182, 41–51. [Google Scholar] [CrossRef]
  6. Li, D.; Dharmawan, D.A.; Ng, B.P.; Rahardja, S. Residual U-Net for Retinal Vessel Segmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1425–1429. [Google Scholar] [CrossRef]
  7. Beheshti, N.; Johnsson, L. Squeeze U-Net: A Memory and Energy Efficient Image Segmentation Network. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1495–1504. [Google Scholar] [CrossRef]
  8. Siddique, N.; Sidike, P.; Reyes, A.; Alom, M.Z.; Devabhaktuni, V. Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation. J. Med. Imaging 2022, 9, 064004. [Google Scholar] [CrossRef]
  9. Mu, N.; Lyu, Z.; Rezaeitaleshmahalleh, M.; Tang, J.; Jiang, J. An attention residual u-net with differential preprocessing and geometric postprocessing: Learning how to segment vasculature including intracranial aneurysms. Med. Image Anal. 2023, 84, 102697. [Google Scholar] [CrossRef]
  10. Chen, J.; Mei, J.; Li, X.; Lu, Y.; Yu, Q.; Wei, Q.; Luo, X.; Xie, Y.; Adeli, E.; Wang, Y.; et al. TransUNet: Rethinking the U-Net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 2024, 97, 103280. [Google Scholar] [CrossRef]
  11. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. arXiv 2021, arXiv:2105.05537. [Google Scholar] [CrossRef]
  12. Valanarasu, J.M.J.; Oza, P.; Hacihaliloglu, I.; Patel, V.M. Medical Transformer: Gated Axial-Attention for Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part I; Springer: Cham, Switzerland, 2021; pp. 36–46. [Google Scholar]
  13. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6877–6886. [Google Scholar] [CrossRef]
  14. Hatamizadeh, A.; Yang, D.; Roth, H.R.; Xu, D. UNETR: Transformers for 3D Medical Image Segmentation. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 1748–1758. [Google Scholar]
  15. Ranftl, R.; Bochkovskiy, A.; Koltun, V. Vision Transformers for Dense Prediction. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 12159–12168. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Liu, H.; Hu, Q. TransFuse: Fusing Transformers and CNNs for Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part I; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer: Cham, Switzerland, 2021; pp. 14–24. [Google Scholar]
  17. Cheng, B.; Schwing, A.G.; Kirillov, A. Per-Pixel Classification is Not All You Need for Semantic Segmentation. Adv. Neural Inf. Process. Syst. 2021, 34, 17864–17875. [Google Scholar]
  18. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. arXiv 2021, arXiv:2105.05633. [Google Scholar] [CrossRef]
  19. Chen, B.; Liu, Y.; Zhang, Z.; Lu, G.; Kong, A.W.K. TransAttUnet: Multi-Level Attention-Guided U-Net with Transformer for Medical Image Segmentation. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 55–68. [Google Scholar] [CrossRef]
  20. Donnan, G.; Baron, J.C.; Davis, S.; Sharp, F. The ischemic penumbra: Overview, definition, and criteria. In The Ischemic Penumbra; CRC Press: Boca Raton, FL, USA, 2007; pp. 7–20. [Google Scholar]
  21. Nael, K.; Tadayon, E.; Wheelwright, D.; Metry, A.; Fifi, J.; Tuhrim, S.; De Leacy, R.; Doshi, A.; Chang, H.; Mocco, J. Defining Ischemic Core in Acute Ischemic Stroke Using CT Perfusion: A Multiparametric Bayesian-Based Model. Am. J. Neuroradiol. 2019, 40, 1491–1497. [Google Scholar] [CrossRef]
  22. Yoo, R.E.; Choi, S.H. Deep learning-based image enhancement techniques for fast MRI in neuroimaging. Magn. Reson. Med. Sci. 2024, 23, 341–351. [Google Scholar] [CrossRef] [PubMed]
  23. Ye, Z.; Luo, S.; Wang, L. Deep Learning Based Cystoscopy Image Enhancement. J. Endourol. 2024, 38, 962–968. [Google Scholar] [CrossRef] [PubMed]
  24. Barrett, J.; Keat, N. Artifacts in CT: Recognition and Avoidance. Radiographics 2004, 24, 1679–1691. [Google Scholar] [CrossRef]
  25. Al-Shakhrah, I.; Al-Obaidi, T. Common artifacts in computerized tomography: A review. Appl. Radiol. 2003, 32, 25–30. [Google Scholar]
  26. Zohair, A.A.; Shamil, A.A.; Sulong, G. Latest methods of image enhancement and restoration for computed tomography: A concise review. Appl. Med. Inform. 2015, 36, 1–12. [Google Scholar]
  27. Park, J.; Lee, J.Y.; Yoo, D.; Kweon, I.S. Distort-and-recover: Color enhancement using deep reinforcement learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5928–5936. [Google Scholar]
  28. Chen, Y.S.; Wang, Y.C.; Kao, M.H.; Chuang, Y.Y. Deep photo enhancer: Unpaired learning for image enhancement from photographs with gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6306–6314. [Google Scholar]
  29. Hu, Y.; He, H.; Xu, C.; Wang, B.; Lin, S. Exposure: A white-box photo post-processing framework. ACM Trans. Graph. (TOG) 2018, 37, 1–17. [Google Scholar] [CrossRef]
  30. Shan, C.; Zhang, Z.; Chen, Z. A coarse-to-fine framework for learned color enhancement with non-local attention. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 949–953. [Google Scholar]
  31. Yang, L.; Yao, S.; Chen, P.; Shen, M.; Fu, S.; Xing, J.; Xue, Y.; Chen, X.; Wen, X.; Zhao, Y.; et al. Unpaired fundus image enhancement based on constrained generative adversarial networks. J. Biophotonics 2024, 17, e202400168. [Google Scholar] [CrossRef]
  32. Tatanov, O.; Samarin, A. LFIEM: Lightweight Filter-based Image Enhancement Model. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 873–878. [Google Scholar]
  33. Samarin, A.; Nazarenko, A.; Savelev, A.; Toropov, A.; Dzestelova, A.; Mikhailova, E.; Motyko, A.; Malykh, V. A Model Based on Universal Filters for Image Color Correction. Pattern Recognit. Image Anal. 2024, 34, 844–854. [Google Scholar] [CrossRef]
  34. Samarin, A.; Nazarenko, A.; Toropov, A.; Kotenko, E.; Dzestelova, A.; Mikhailova, E.; Malykh, V.; Savelev, A.; Motyko, A. Universal Filter-Based Lightweight Image Enhancement Model with Unpaired Learning Mode. In Proceedings of the 2024 36th Conference of Open Innovations Association (FRUCT), Lappeenranta, Finland, 30 October–1 November 2024; pp. 711–720. [Google Scholar]
  35. Kosugi, S.; Yamasaki, T. Unpaired Image Enhancement Featuring Reinforcement-Learning-Controlled Image Editing Software. Proc. AAAI Conf. Artif. Intell. 2020, 34, 11296–11303. [Google Scholar] [CrossRef]
  36. Wang, Y.; Xu, T.; Fan, Z.; Xue, T.; Gu, J. AdaptiveISP: Learning an Adaptive Image Signal Processor for Object Detection. In Advances in Neural Information Processing Systems 37 (NeurIPS 2024); Globerson, A., Mackey, L., Belgrave, D., Fan, A., Paquet, U., Tomczak, J., Zhang, C., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2024; Volume 37, pp. 112598–112623. [Google Scholar]
  37. Umerenkov, D.; Kudin, S.; Peksheva, M.; Pavlov, D. CPAISD: Core-penumbra acute ischemic stroke dataset. arXiv 2024, arXiv:2404.02518. [Google Scholar] [CrossRef]
  38. Mohamed, A.; Rabea, M.; Sameh, A.; Kamal, E. Brain Tumor Radiogenomic Classification. arXiv 2024, arXiv:2401.09471. [Google Scholar] [CrossRef]
  39. Chilamkurthy, S.; Ghosh, R.; Tanamala, S.; Biviji, M.; Campeau, N.G.; Venugopal, V.K.; Mahajan, V.; Rao, P.; Warier, P. Development and Validation of Deep Learning Algorithms for Detection of Critical Findings in Head CT Scans. arXiv 2018, arXiv:1803.05854. [Google Scholar] [CrossRef]
  40. Wang, Y.; Xu, T.; Fan, Z.; Xue, T.; Gu, J. AdaptiveISP: Learning an Adaptive Image Signal Processor for Object Detection. Adv. Neural Inf. Process. Syst. 2025, 37, 112598–112623. [Google Scholar]
  41. He, J.; Liu, Y.; Qiao, Y.; Dong, C. Conditional sequential modulation for efficient global image retouching. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIII 16; Springer: Cham, Switzerland, 2020; pp. 679–695. [Google Scholar]
  42. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.I.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
  43. Narayan, S. The generalized sigmoid activation function: Competitive supervised learning. Inf. Sci. 1997, 99, 69–82. [Google Scholar] [CrossRef]
  44. Adnan, F.H.; Mahmud, M.F.O.; Abdullah, W.F.H. Hyperbolic tangent activation function integrated circuit implementation for perceptrons. In Proceedings of the 2012 IEEE Student Conference on Research and Development (SCOReD), Pulau Pinang, Malaysia, 5–6 December 2012; pp. 84–87. [Google Scholar] [CrossRef]
  45. Apicella, A.; Donnarumma, F.; Isgrò, F.; Prevete, R. A survey on modern trainable activation functions. Neural Netw. 2020, 138, 14–32. [Google Scholar] [CrossRef]
  46. UmaMaheswaran, S.; Ahmad, F.; Hegde, R.; Alwakeel, A.M.; Rameem Zahra, S. Enhanced non-contrast computed tomography images for early acute stroke detection using machine learning approach. Expert Syst. Appl. 2024, 240, 122559. [Google Scholar] [CrossRef]
  47. Hayat, M.; Gupta, M.; Suanpang, P.; Nanthaamornphong, A. Super-Resolution Methods for Endoscopic Imaging: A Review. In Proceedings of the 2024 12th International Conference on Internet of Everything, Microwave, Embedded, Communication and Networks (IEMECON), Jaipur, India, 24–26 October 2024; pp. 1–6. [Google Scholar] [CrossRef]
  48. Ding, F.; Shi, Y.; Zhu, G.; Shi, Y. Real-time estimation for the parameters of Gaussian filtering via deep learning. J. Real-Time Image Process. 2020, 17, 17–27. [Google Scholar] [CrossRef]
  49. Chernov, A. On using Gaussian functions with varied parameters for approximation of functions of one variable on a finite segment. Vestn. Udmurt. Univ. Mat. Mekhanika Komp’yuternye Nauk. 2017, 27, 267–282. [Google Scholar] [CrossRef]
  50. Lee, D.; In, J.; Lee, S. Standard deviation and standard error of the mean. Korean J. Anesthesiol. 2015, 68, 220–223. [Google Scholar] [CrossRef]
  51. Villar, S.; Torcida, S.; Acosta, G. Median Filtering: A New Insight. J. Math. Imaging Vis. 2017, 58, 130–146. [Google Scholar] [CrossRef]
  52. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar] [CrossRef]
  53. Britto, A.; Chinnasamy, M. Deep learning with fine tuning strategy for Image segmentation and classification. J. Xidian Univ. 2023, 14, 1107–1113. [Google Scholar] [CrossRef]
  54. Xie, W.; Willems, N.; Patil, S.; Li, Y.; Kumar, M. SAM Fewshot Finetuning for Anatomical Segmentation in Medical Images. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 3241–3249. [Google Scholar] [CrossRef]
  55. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  56. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
  57. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  58. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  59. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180. [Google Scholar] [CrossRef]
  60. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  61. Zhao, R.; Qian, B.; Zhang, X.; Li, Y.; Wei, R.; Liu, Y.; Pan, Y. Rethinking Dice Loss for Medical Image Segmentation. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 851–860. [Google Scholar] [CrossRef]
  62. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the Third International Workshop, DLMIA 2017, and 7th International Workshop, ML-CDS 2017, Held in Conjunction with MICCAI 2017, Québec City, QC, Canada, 14 September 2017; Proceedings; Cardoso, M.J., Arbel, T., Carneiro, G., Syeda-Mahmood, T., Tavares, J.M.R., Moradi, M., Bradley, A., Greenspan, H., Papa, J.P., Madabhushi, A., et al., Eds.; Springer: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar]
  63. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  64. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  65. Kiware Medical Dicom Anonymizer. 2024. Available online: https://github.com/KitwareMedical/dicom-anonymizer (accessed on 21 June 2025).
  66. Mumuni, A.; Mumuni, F. Data augmentation: A comprehensive survey of modern approaches. Array 2022, 16, 100258. [Google Scholar] [CrossRef]
  67. Shorten, C.; Khoshgoftaar, T. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  68. Sheng, H.; Cai, S.; Zhao, N.; Deng, B.; Huang, J.; Hua, X.S.; Zhao, M.J.; Lee, G. Rethinking IoU-based Optimization for Single-stage 3D Object Detection. In Computer Vision—ECCV 2022, Proceedings of the 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part IX; Springer: Cham, Switzerland, 2022; pp. 544–561. [Google Scholar] [CrossRef]
  69. Wang, H.; Cong, Y.; Litany, O.; Gao, Y.; Guibas, L. 3DIoUMatch: Leveraging IoU Prediction for Semi-Supervised 3D Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14610–14619. [Google Scholar] [CrossRef]
  70. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and A Loss for Bounding Box Regression. arXiv 2019, arXiv:1902.09630. [Google Scholar] [CrossRef]
  71. Stodt, J.; Reich, C.; Clarke, N. Unified Intersection Over Union for Explainable Artificial Intelligence. In Intelligent Systems and Applications, Proceedings of the 2023 Intelligent Systems Conference (IntelliSys) Volume 2, Amsterdam, The Netherlands, 7–8 September 2024; Springer: Cham, Switzerland, 2024; pp. 758–770. [Google Scholar] [CrossRef]
  72. Taha, A.A.; Hanbury, A. Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef]
  73. Müller, D.; Soto-Rey, I.; Kramer, F. Towards a guideline for evaluation metrics in medical image segmentation. BMC Res. Notes 2022, 15, 210. [Google Scholar] [CrossRef]
  74. Bertels, J.; Eelbode, T.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimizing the Dice Score and Jaccard Index for Medical Image Segmentation: Theory & Practice. arXiv 2019, arXiv:1911.01685. [Google Scholar] [CrossRef]
  75. Rainio, O.; Klén, R. Modified Dice Coefficients for Evaluation of Tumor Segmentation from PET Images: A Proof-of-Concept Study. J. Imaging Inform. Med. 2025, 1–9. [Google Scholar] [CrossRef] [PubMed]
  76. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  77. Parmar, J.; Satheesh, S.; Patwary, M.; Shoeybi, M.; Catanzaro, B. Reuse, Don’t Retrain: A Recipe for Continued Pretraining of Language Models. arXiv 2024, arXiv:2407.07263. [Google Scholar] [CrossRef]
  78. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar] [CrossRef]
  79. Tian, Q.; Wang, Z.; Cui, X. Improved Unet brain tumor image segmentation based on GSConv module and ECA attention mechanism. Appl. Comput. Eng. 2024, 88, 214–223. [Google Scholar] [CrossRef]
  80. Larson, P.B.; Zapletal, J. Discontinuous homomorphisms, selectors and automorphisms of the complex field. arXiv 2020, arXiv:1803.02740. [Google Scholar] [CrossRef]
  81. Lauande, M.G.M.; Braz Junior, G.; de Almeida, J.D.S.; Silva, A.C.; Gil da Costa, R.M.; Teles, A.M.; da Silva, L.L.; Brito, H.O.; Vidal, F.C.B.; do Vale, J.G.A.; et al. Building a DenseNet-Based Neural Network with Transformer and MBConv Blocks for Penile Cancer Classification. Appl. Sci. 2024, 14, 10536. [Google Scholar] [CrossRef]
  82. Rai, K.; Hojatpanah, F.; Ajaei, F.; Grolinger, K. Deep Learning for High-Impedance Fault Detection: Convolutional Autoencoders. Energies 2021, 14, 3623. [Google Scholar] [CrossRef]
  83. Samarin, A.; Toropov, A.; Egorova, O. Self-Attention Based Approach to Iris Segmentation. In Proceedings of the 2025 International Russian Smart Industry Conference (SmartIndustryCon), Sochi, Russia, 24–28 March 2025; pp. 200–205. [Google Scholar] [CrossRef]
  84. Samarin, A.; Savelev, A.; Toropov, A.; Dzestelova, A.; Malykh, V.; Mikhailova, E.; Motyko, A. Prior Segmentation and Attention Based Approach to Neoplasms Recognition by Single-Channel Monochrome Computer Tomography Snapshots. In In Pattern Recognition, Computer Vision, and Image Processing, Proceedings of the ICPR 2022 International Workshops and Challenges, Montreal, QC, Canada, 21–25 August 2022; Proceedings, Part II; Springer: Berlin/Heidelberg, Germany, 2022; pp. 561–570. [Google Scholar] [CrossRef]
  85. Rais, K.; Amroune, M.; Benmachiche, A.; Haouam, M.Y. Exploring Variational Autoencoders for Medical Image Generation: A Comprehensive Study. arXiv 2024, arXiv:2411.07348. [Google Scholar] [CrossRef]
  86. Jha, D.; Riegler, M.A.; Johansen, D.; Halvorsen, P.; Johansen, H.D. DoubleU-Net: A Deep Convolutional Neural Network for Medical Image Segmentation. arXiv 2020, arXiv:2006.04868. [Google Scholar] [CrossRef]
  87. Jose, J.M.; Sindagi, V.; Hacihaliloglu, I.; Patel, V.M. KiU-Net: Towards Accurate Segmentation of Biomedical Images using Over-complete Representations. arXiv 2020, arXiv:2006.04878. [Google Scholar] [CrossRef]
  88. Gui, H.; Wang, R.; Yin, K.; Jin, L.; Kula, M.; Xu, T.; Hong, L.; Chi, E.H. Hiformer: Heterogeneous Feature Interactions Learning with Transformers for Recommender Systems. arXiv 2023, arXiv:2311.05884. [Google Scholar] [CrossRef]
Figure 1. An example of an input image with three classes. The first one is the original CT scan, the second is the ischemic core mask, the third is the penumbra area mask, and the fourth is the visualization with mask overlay (ischemic core in red, penumbra area in yellow).
Figure 1. An example of an input image with three classes. The first one is the original CT scan, the second is the ischemic core mask, the third is the penumbra area mask, and the fourth is the visualization with mask overlay (ischemic core in red, penumbra area in yellow).
Jimaging 11 00359 g001
Figure 2. The conceptual scheme of the two-stage solution we proposed for segmenting NCCT scans of the human brain.
Figure 2. The conceptual scheme of the two-stage solution we proposed for segmenting NCCT scans of the human brain.
Jimaging 11 00359 g002
Figure 3. Architecture of the proposed lightweight image preprocessing module.
Figure 3. Architecture of the proposed lightweight image preprocessing module.
Jimaging 11 00359 g003
Figure 4. UNet encoder-decoder based kernels pretraining scheme.
Figure 4. UNet encoder-decoder based kernels pretraining scheme.
Jimaging 11 00359 g004
Figure 5. Examples of annotated images from the CPAISD dataset. The first and third columns present the original NCCT scans, while the second and fourth columns display the masked regions. The ischemic core is shown in red, and the penumbra is shown in yellow.
Figure 5. Examples of annotated images from the CPAISD dataset. The first and third columns present the original NCCT scans, while the second and fourth columns display the masked regions. The ischemic core is shown in red, and the penumbra is shown in yellow.
Jimaging 11 00359 g005
Figure 6. Example of the applied “soft” augmentations to the original image to create a paired dataset. The first one is the original image; all others are modified images.
Figure 6. Example of the applied “soft” augmentations to the original image to create a paired dataset. The first one is the original image; all others are modified images.
Jimaging 11 00359 g006
Figure 7. An example of the baseline model’s performance with and without our preprocessing module is shown. Colors: yellow—penumbra region detected by the model without preprocessing; purple—penumbra region detected by the model with preprocessing; light green—ground truth penumbra area. In the right image, the ischemic core is also detected. For better visualization of the ground truth ischemic core annotation and the core obtained using our model, we have added a black contour around it. Colors: red—ischemic core detected using the preprocessing model; light purple—ground truth ischemic core annotation; green—ischemic core detected by the model without preprocessing.
Figure 7. An example of the baseline model’s performance with and without our preprocessing module is shown. Colors: yellow—penumbra region detected by the model without preprocessing; purple—penumbra region detected by the model with preprocessing; light green—ground truth penumbra area. In the right image, the ischemic core is also detected. For better visualization of the ground truth ischemic core annotation and the core obtained using our model, we have added a black contour around it. Colors: red—ischemic core detected using the preprocessing model; light purple—ground truth ischemic core annotation; green—ischemic core detected by the model without preprocessing.
Jimaging 11 00359 g007
Table 1. Top 30 autoencoder configurations with varied convolutional skip connections (kernel combinations shown in parentheses). All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Table 1. Top 30 autoencoder configurations with varied convolutional skip connections (kernel combinations shown in parentheses). All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
RankBackboneDepthSkip KernelsParams (M)Dice (P)Dice (C)mIoU (P)mIoU (C)
1U-Net [79]55 × 5 + 7 × 7 + 11 × 1134.20.6180.6020.5010.487
2U-Net53 × 3 + 5 × 5 + 9 × 934.00.6150.5980.4970.483
3ResAE-U-Net [80]55 × 5 + 7 × 7 + 11 × 1136.70.6130.5960.4950.481
4U-Net43 × 3 + 5 × 5 + 7 × 722.90.6110.5940.4930.479
5U-Net + SE55 × 5 + 9 × 935.30.6090.5920.4910.477
6DenseAE-U-Net [81]53 × 3 + 7 × 7 + 11 × 1140.00.6070.5900.4890.475
7U-Net53 × 3 + 5 × 533.80.6050.5880.4870.473
8U-Net45 × 5 + 7 × 722.80.6030.5860.4850.471
9CAE (sym) [82]55 × 5 + 7 × 7 + 11 × 1128.50.6010.5840.4830.469
10U-Net33 × 3 + 5 × 5 + 7 × 712.60.5990.5820.4810.467
11VAE57 × 7 + 11 × 1129.30.5970.5800.4790.465
12U-Net + Attn [79,83,84]53 × 3 + 5 × 5 + 9 × 936.10.5950.5780.4770.463
13ResAE (sym)53 × 3 + 7 × 731.70.5930.5760.4750.461
14U-Net57 × 7 only33.70.5910.5740.4730.459
15DenseAE (sym)55 × 5 + 9 × 933.80.5890.5720.4710.457
16U-Net43 × 3 + 11 × 1122.70.5870.5700.4690.455
17CAE (asym)55 × 5 + 7 × 726.90.5850.5680.4670.453
18U-Net + GN43 × 3 + 5 × 5 + 9 × 922.90.5830.5660.4650.451
19VAE + L155 × 5 + 7 × 729.30.5810.5640.4630.449
20ResAE + SE53 × 3 + 5 × 5 + 7 × 732.20.5790.5620.4610.447
21U-Net511 × 11 only34.10.5770.5600.4590.445
22CAE + SpecNorm53 × 3 + 5 × 528.50.5750.5580.4570.443
23U-Net + Drop55 × 5 + 9 × 934.40.5730.5560.4550.441
24DenseAE + Drop53 × 3 + 7 × 733.80.5710.5540.4530.439
25ResAE (shallow)45 × 5 + 7 × 721.50.5690.5520.4510.437
26CAE (wide)43 × 3 + 5 × 5 + 7 × 726.10.5670.5500.4490.435
27VAE + β = 0.5 [85]53 × 3 + 5 × 529.30.5650.5480.4470.433
28U-Net + L257 × 7 + 9 × 934.40.5630.5460.4450.431
29CAE + ELU55 × 5 only28.50.5610.5440.4430.429
30ResAE + ELU53 × 3 + 5 × 531.70.5590.5420.4410.427
Table 2. Top 50 architectures for stroke lesion segmentation on CPAISD comparing CP (custom preprocessing), PP (standard preprocessing), and Base (no preprocessing) approaches. Metrics show Dice and mIoU for penumbra (P) and core (C) regions in 2D/3D. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Table 2. Top 50 architectures for stroke lesion segmentation on CPAISD comparing CP (custom preprocessing), PP (standard preprocessing), and Base (no preprocessing) approaches. Metrics show Dice and mIoU for penumbra (P) and core (C) regions in 2D/3D. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
RankTypeModelDice 2D (P/C)mIoU 2D (P/C)Dice 3D (P/C)mIoU 3D (P/C)
1CPUNet++0.624/0.6080.507/0.4930.608/0.5920.491/0.477
2CPSwinUNet0.623/0.6070.506/0.4920.607/0.5910.490/0.476
3CPDoubleUNet0.619/0.6030.502/0.4880.605/0.5890.488/0.474
4CPTransUNet0.618/0.6020.501/0.4870.604/0.5880.487/0.473
5PPnnUNet0.620/0.6040.503/0.4890.603/0.5870.486/0.472
6CPAttnUNet0.616/0.6000.499/0.4850.602/0.5860.485/0.471
7CPResUNet++0.615/0.5990.498/0.4840.601/0.5850.484/0.470
8CPMultiResUNet0.614/0.5980.497/0.4830.600/0.5840.483/0.469
9PPUNet3+0.615/0.5990.498/0.4840.599/0.5830.482/0.468
10CPKiU-Net0.613/0.5970.496/0.4820.598/0.5820.481/0.467
11CPFocalUNet0.612/0.5960.495/0.4810.597/0.5810.480/0.466
12PPCE-Net0.612/0.5960.495/0.4810.596/0.5800.479/0.465
13CPDenseUNet0.611/0.5950.494/0.4800.596/0.5800.479/0.465
14CPInf-Net0.610/0.5940.493/0.4790.595/0.5790.478/0.464
15CPPraNet0.609/0.5930.492/0.4780.594/0.5780.477/0.463
16BaseUCTransNet0.605/0.5890.488/0.4740.590/0.5740.473/0.459
17PPMedT0.606/0.5900.489/0.4750.591/0.5750.474/0.460
18BaseMISSU0.603/0.5870.486/0.4720.588/0.5720.471/0.457
19PPCrossFormer0.604/0.5880.487/0.4730.589/0.5730.472/0.458
20BaseBTSNet0.601/0.5850.484/0.4700.586/0.5700.469/0.455
21CPCS2-Net0.608/0.5920.491/0.4770.593/0.5770.476/0.462
22CPHiFormer0.607/0.5910.490/0.4760.592/0.5760.475/0.461
23PPFAT-Net0.605/0.5890.488/0.4740.590/0.5740.473/0.459
24BaseDC-UNet0.602/0.5860.485/0.4710.587/0.5710.470/0.456
25CPColonSegNet0.604/0.5880.487/0.4730.589/0.5730.472/0.458
26PPTransClaw U-Net0.603/0.5870.486/0.4720.588/0.5720.471/0.457
27BaseMHSA-UNet0.600/0.5840.483/0.4690.585/0.5690.468/0.454
28CPD2A-UNet0.601/0.5850.484/0.4700.586/0.5700.469/0.455
29PPSCUNet0.599/0.5830.482/0.4680.584/0.5680.467/0.453
30BaseFSS-UNet0.597/0.5810.480/0.4660.582/0.5660.465/0.451
31CP3D-UNet0.598/0.5820.481/0.4670.583/0.5670.466/0.452
32PPUNeXt0.596/0.5800.479/0.4650.581/0.5650.464/0.450
33BasePVT-UNet0.594/0.5780.477/0.4630.579/0.5630.462/0.448
34CPCoBiNet0.595/0.5790.478/0.4640.580/0.5640.463/0.449
35PPEdge-UNet0.593/0.5770.476/0.4620.578/0.5620.461/0.447
36BaseHFA-UNet0.591/0.5750.474/0.4600.576/0.5600.459/0.445
37CPTDB-UNet0.592/0.5760.475/0.4610.577/0.5610.460/0.446
38PPMCGU-Net0.590/0.5740.473/0.4590.575/0.5590.458/0.444
39BaseSCPM-Net0.588/0.5720.471/0.4570.573/0.5570.456/0.442
40CPDPR-UNet0.589/0.5730.472/0.4580.574/0.5580.457/0.443
41PPCFM-UNet0.587/0.5710.470/0.4560.572/0.5560.455/0.441
42BaseLGANet0.585/0.5690.468/0.4540.570/0.5540.453/0.439
43CPVM-UNet0.586/0.5700.469/0.4550.571/0.5550.454/0.440
44PPDCR-UNet0.584/0.5680.467/0.4530.569/0.5530.452/0.438
45BaseGSU-Net0.582/0.5660.465/0.4510.567/0.5510.450/0.436
46CPFRT-UNet0.583/0.5670.466/0.4520.568/0.5520.451/0.437
47PPADA-UNet0.581/0.5650.464/0.4500.566/0.5500.449/0.435
48BasePCAC-UNet0.579/0.5630.462/0.4480.564/0.5480.447/0.433
49CPSparse-UNet0.580/0.5640.463/0.4490.565/0.5490.448/0.434
50PPDSR-UNet0.578/0.5620.461/0.4470.563/0.5470.446/0.432
Table 3. Top 50 architectures for stroke lesion segmentation on RSNA-MICCAI + CQ500 comparing CP (custom preprocessing), PP (standard preprocessing), and Base approaches. Metrics show Dice and mIoU for penumbra (P) and core (C) regions in 2D/3D. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Table 3. Top 50 architectures for stroke lesion segmentation on RSNA-MICCAI + CQ500 comparing CP (custom preprocessing), PP (standard preprocessing), and Base approaches. Metrics show Dice and mIoU for penumbra (P) and core (C) regions in 2D/3D. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
RankTypeModelDice 2D (P/C)mIoU 2D (P/C)Dice 3D (P/C)mIoU 3D (P/C)
1CPSwinUNet0.712/0.6980.602/0.5880.695/0.6810.585/0.571
2CPDoubleUNet0.710/0.6960.600/0.5860.693/0.6790.583/0.569
3PPnnUNet0.709/0.6950.599/0.5850.692/0.6780.582/0.568
4CPUNet++0.708/0.6940.598/0.5840.691/0.6770.581/0.567
5CPTransUNet0.707/0.6930.597/0.5830.690/0.6760.580/0.566
6PPUNet3+0.706/0.6920.596/0.5820.689/0.6750.579/0.565
7CPAttnUNet0.705/0.6910.595/0.5810.688/0.6740.578/0.564
8CPResUNet++0.704/0.6900.594/0.5800.687/0.6730.577/0.563
9PPCE-Net0.703/0.6890.593/0.5790.686/0.6720.576/0.562
10CPMultiResUNet0.702/0.6880.592/0.5780.685/0.6710.575/0.561
11CPKiU-Net0.701/0.6870.591/0.5770.684/0.6700.574/0.560
12PPMedT0.700/0.6860.590/0.5760.683/0.6690.573/0.559
13CPFocalUNet0.699/0.6850.589/0.5750.682/0.6680.572/0.558
14CPDenseUNet0.698/0.6840.588/0.5740.681/0.6670.571/0.557
15PPCrossFormer0.697/0.6830.587/0.5730.680/0.6660.570/0.556
16CPInf-Net0.696/0.6820.586/0.5720.679/0.6650.569/0.555
17BaseUCTransNet0.692/0.6780.582/0.5680.675/0.6610.565/0.551
18CPPraNet0.695/0.6810.585/0.5710.678/0.6640.568/0.554
19PPFAT-Net0.694/0.6800.584/0.5700.677/0.6630.567/0.553
20BaseBTSNet0.690/0.6760.580/0.5660.673/0.6590.563/0.549
21CPCS2-Net0.693/0.6790.583/0.5690.676/0.6620.566/0.552
22BaseMISSU0.689/0.6750.579/0.5650.672/0.6580.562/0.548
23CPHiFormer0.692/0.6780.582/0.5680.675/0.6610.565/0.551
24PPTransClaw U-Net0.691/0.6770.581/0.5670.674/0.6600.564/0.550
25BaseDC-UNet0.688/0.6740.578/0.5640.671/0.6570.561/0.547
26CPColonSegNet0.690/0.6760.580/0.5660.673/0.6590.563/0.549
27PPSCUNet0.687/0.6730.577/0.5630.670/0.6560.560/0.546
28BaseMHSA-UNet0.686/0.6720.576/0.5620.669/0.6550.559/0.545
29CPD2A-UNet0.689/0.6750.579/0.5650.672/0.6580.562/0.548
30BaseFSS-UNet0.685/0.6710.575/0.5610.668/0.6540.558/0.544
31PPUNeXt0.684/0.6700.574/0.5600.667/0.6530.557/0.543
32CP3D-UNet0.688/0.6740.578/0.5640.671/0.6570.561/0.547
33BasePVT-UNet0.683/0.6690.573/0.5590.666/0.6520.556/0.542
34PPEdge-UNet0.682/0.6680.572/0.5580.665/0.6510.555/0.541
35CPCoBiNet0.687/0.6730.577/0.5630.670/0.6560.560/0.546
36BaseHFA-UNet0.681/0.6670.571/0.5570.664/0.6500.554/0.540
37PPMCGU-Net0.680/0.6660.570/0.5560.663/0.6490.553/0.539
38CPTDB-UNet0.686/0.6720.576/0.5620.669/0.6550.559/0.545
39BaseSCPM-Net0.679/0.6650.569/0.5550.662/0.6480.552/0.538
40PPCFM-UNet0.678/0.6640.568/0.5540.661/0.6470.551/0.537
41CPDPR-UNet0.685/0.6710.575/0.5610.668/0.6540.558/0.544
42BaseLGANet0.677/0.6630.567/0.5530.660/0.6460.550/0.536
43PPDCR-UNet0.676/0.6620.566/0.5520.659/0.6450.549/0.535
44CPVM-UNet0.684/0.6700.574/0.5600.667/0.6530.557/0.543
45BaseGSU-Net0.675/0.6610.565/0.5510.658/0.6440.548/0.534
46PPADA-UNet0.674/0.6600.564/0.5500.657/0.6430.547/0.533
47CPFRT-UNet0.683/0.6690.573/0.5590.666/0.6520.556/0.542
48BasePCAC-UNet0.673/0.6590.563/0.5490.656/0.6420.546/0.532
49PPDSR-UNet0.672/0.6580.562/0.5480.655/0.6410.545/0.531
50CPSparse-UNet0.682/0.6680.572/0.5580.665/0.6510.555/0.541
Table 4. Comprehensive evaluation of preprocessing filter combinations for CP and SwinUNet architecture on CPAISD+RSNA-MICCAI+CQ500 mixed dataset. Metrics show Dice and mIoU scores for penumbra (P) and core (C) regions in both 2D and 3D evaluations. CB stands for custom block. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Table 4. Comprehensive evaluation of preprocessing filter combinations for CP and SwinUNet architecture on CPAISD+RSNA-MICCAI+CQ500 mixed dataset. Metrics show Dice and mIoU scores for penumbra (P) and core (C) regions in both 2D and 3D evaluations. CB stands for custom block. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Preprocessing CombinationDice 2D (P)Dice 2D (C)Dice 3D (P)Dice 3D (C)mIoU 2D (P)mIoU 2D (C)mIoU 3D (P)mIoU 3D (C)
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + sharpen + contrast0.6840.6700.6670.6530.5740.5600.5570.543
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + contrast + laplacian0.6810.6670.6640.6500.5710.5570.5540.540
CB 5 × 5 + 7 × 7 + 11 × 11 + sharpen + contrast + exposure0.6820.6680.6650.6510.5720.5580.5550.541
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + sharpen + trainable kernel0.6830.6690.6660.6520.5730.5590.5560.542
CB 5 × 5 + 7 × 7 + 11 × 11 + linear + sharpen + contrast0.6800.6660.6630.6490.5700.5560.5530.539
brightness + contrast + sharpen + laplacian0.6750.6610.6580.6440.5650.5510.5480.534
blur + sharpen + linear + exposure0.6720.6580.6550.6410.5620.5480.5450.531
laplacian + sharpen + linear + trainable kernel0.6730.6590.6560.6420.5630.5490.5460.532
gaussian blur + median filter + sharpen0.6700.6560.6530.6390.5600.5460.5430.529
bilateral filter + contrast enhancement0.6680.6540.6510.6370.5580.5440.5410.527
universal kernel + linear0.6690.6550.6520.6380.5590.5450.5420.528
brightness + linear0.6670.6530.6500.6360.5570.5430.5400.526
Table 5. Comprehensive evaluation of preprocessing filter combinations for CP and UNet++ architecture on CPAISD+RSNA-MICCAI+CQ500 mixed dataset. Metrics show Dice and mIoU scores for penumbra (P) and core (C) regions in both 2D and 3D evaluations. CB stands for custom block. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Table 5. Comprehensive evaluation of preprocessing filter combinations for CP and UNet++ architecture on CPAISD+RSNA-MICCAI+CQ500 mixed dataset. Metrics show Dice and mIoU scores for penumbra (P) and core (C) regions in both 2D and 3D evaluations. CB stands for custom block. All measurements presented were obtained across 10 random seeds: Mean ± 95% CI with the highest observed variance < 0.001.
Preprocessing CombinationDice 2D (P)Dice 2D (C)Dice 3D (P)Dice 3D (C)mIoU 2D (P)mIoU 2D (C)mIoU 3D (P)mIoU 3D (C)
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + sharpen + contrast0.6820.6680.6650.6510.5720.5580.5550.541
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + contrast + laplacian0.6790.6650.6620.6480.5690.5550.5520.538
CB 5 × 5 + 7 × 7 + 11 × 11 + sharpen + contrast + exposure0.6800.6660.6630.6490.5700.5560.5530.539
CB 5 × 5 + 7 × 7 + 11 × 11 + brightness + sharpen + trainable kernel0.6810.6670.6640.6500.5710.5570.5540.540
CB 5 × 5 + 7 × 7 + 11 × 11 + linear + sharpen + contrast0.6780.6640.6610.6470.5680.5540.5510.537
brightness + contrast + sharpen + laplacian0.6730.6590.6560.6420.5630.5490.5460.532
blur + sharpen + linear + exposure0.6700.6560.6530.6390.5600.5460.5430.529
laplacian + sharpen + linear + trainable kernel0.6710.6570.6540.6400.5610.5470.5440.530
gaussian blur + median filter + sharpen0.6680.6540.6510.6370.5580.5440.5410.527
brightness + contrast enhancement0.6660.6520.6490.6350.5560.5420.5390.525
contrast + linear0.6670.6530.6500.6360.5570.5430.5400.526
universal kernel + linear0.6650.6510.6480.6340.5550.5410.5380.524
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Samarin, A.; Savelev, A.; Toropov, A.; Dozortseva, A.; Kotenko, E.; Nazarenko, A.; Motyko, A.; Narova, G.; Mikhailova, E.; Malykh, V. Non-Contrast Brain CT Images Segmentation Enhancement: Lightweight Pre-Processing Model for Ultra-Early Ischemic Lesion Recognition and Segmentation. J. Imaging 2025, 11, 359. https://doi.org/10.3390/jimaging11100359

AMA Style

Samarin A, Savelev A, Toropov A, Dozortseva A, Kotenko E, Nazarenko A, Motyko A, Narova G, Mikhailova E, Malykh V. Non-Contrast Brain CT Images Segmentation Enhancement: Lightweight Pre-Processing Model for Ultra-Early Ischemic Lesion Recognition and Segmentation. Journal of Imaging. 2025; 11(10):359. https://doi.org/10.3390/jimaging11100359

Chicago/Turabian Style

Samarin, Aleksei, Alexander Savelev, Aleksei Toropov, Aleksandra Dozortseva, Egor Kotenko, Artem Nazarenko, Alexander Motyko, Galiya Narova, Elena Mikhailova, and Valentin Malykh. 2025. "Non-Contrast Brain CT Images Segmentation Enhancement: Lightweight Pre-Processing Model for Ultra-Early Ischemic Lesion Recognition and Segmentation" Journal of Imaging 11, no. 10: 359. https://doi.org/10.3390/jimaging11100359

APA Style

Samarin, A., Savelev, A., Toropov, A., Dozortseva, A., Kotenko, E., Nazarenko, A., Motyko, A., Narova, G., Mikhailova, E., & Malykh, V. (2025). Non-Contrast Brain CT Images Segmentation Enhancement: Lightweight Pre-Processing Model for Ultra-Early Ischemic Lesion Recognition and Segmentation. Journal of Imaging, 11(10), 359. https://doi.org/10.3390/jimaging11100359

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop