Next Article in Journal
Energy-Preserving Algorithms for the Ginzburg–Landau Equation: Integrating Scientific Research with Talent Cultivation
Previous Article in Journal
On Fractional Simpson-Type Inequalities via Harmonic Convexity
Previous Article in Special Issue
Int.2D-3D-CNN: Integrated 2D and 3D Convolutional Neural Networks for Video Violence Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Edge-Aware Illumination Enhancement for Fine-Grained Defect Detection on Railway Surfaces

Department of Electrical Engineering, Soonchunhyang University, Asan 31538, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(23), 3780; https://doi.org/10.3390/math13233780
Submission received: 2 October 2025 / Revised: 13 November 2025 / Accepted: 18 November 2025 / Published: 25 November 2025
(This article belongs to the Special Issue Applications of Deep Learning and Convolutional Neural Network)

Abstract

Fine-grained defects on rail surfaces are often inadequately detected by conventional vision-based object detection models in low-light environments. Although this problem can be mitigated by enhancing image brightness and contrast or employing deep learning-based object detectors, these methods frequently distort critical edge and texture information essential for accurate defect recognition. Herein, we propose a preprocessing framework that integrates two complementary modules, namely adaptive illumination enhancement (AIE) and EdgeSeal enhancement (ESE). AIE leverages contrast-limited adaptive histogram equalization and gamma correction to enhance local contrast while adjusting the global brightness distribution. ESE further refines defect visibility through morphological closing and sharpening, enhancing edge continuity and structural clarity. When integrated with the You Only Look Once v11 (YOLOv11) object detection model and evaluated on a rail defect dataset, the proposed framework achieves an ~7% improvement in mean average precision over baseline YOLOv11 and outperforms recent state-of-the-art detectors under diverse low-light and degraded-visibility conditions. The improved precision and recall across three defect classes (defects, dirt, and gaps) demonstrate the robustness of our approach. The proposed framework holds promise for real-time railway infrastructure monitoring and automation systems and is broadly applicable to low-light object detection tasks across other industrial domains.

1. Introduction

Urban railways have become a fundamental transportation mode because of their high operational speed, large passenger capacity, and environmental sustainability. However, structural defects on rail surfaces pose serious threats to operational safety and passenger comfort and can cause system-wide inefficiencies [1,2]. When not detected early, fine-grained defects may accumulate over time, increasing the risk of derailment, mechanical failure, and service disruption, which can result in severe accidents and considerable economic losses. Traditional rail defect inspections heavily rely on manual visual assessments, which are labor-intensive, inconsistent, and prone to human error [3]. Consequently, automated, accurate, and robust rail defect detection systems are highly sought after.
Automated inspection techniques can be broadly classified into physical sensor-based nondestructive testing (NDT) and vision-based approaches. NDT methods, such as ultrasonic, magnetic particle, and eddy current testing, are effective for detecting subsurface defects without physically damaging the infrastructure. For instance, guided wave-based ultrasonic structural health monitoring has been used for real-time rail defect detection [4], and three-dimensional magnetic profiling and Fourier transform-based eddy current analysis have been used for efficient surface anomaly detection [5,6]. Despite their effectiveness, these methods often suffer from high equipment costs and dependency on skilled operators.
In contrast, vision-based methods leveraging deep learning and low-cost imaging devices offer a scalable and efficient alternative [7,8]. Recent efforts have focused on using and optimizing convolutional neural networks, particularly variants of You Only Look Once (YOLO), for rail defect detection under normal lighting conditions [9,10,11]. However, under low-light scenarios common in tunnels, stations, and nighttime environments, these models struggle because of low contrast, reduced texture visibility, and loss of boundary details [12].
Various image enhancement techniques have been introduced as preprocessing steps to improve visual quality prior to detection and thus mitigate the abovementioned problems. Methods such as histogram equalization, Retinex-based enhancement, and generative adversarial network (GAN)-based restoration improve global brightness and contrast [13,14,15] but often fail to preserve the fine contours and textures critical for detecting small defects. Moreover, conventional models frequently lack adaptability to variable illumination conditions, which hinders feature extraction and lowers detection accuracy [16,17,18]. While recent low-light image enhancement methods have achieved notable improvements by combining illumination correction with structural preservation, many of these approaches either require complex network modifications or lack modularity, which limits their applicability to existing detection frameworks.
Herein, we propose a modular two-stage preprocessing framework that enhances low-light images without altering the structure of the downstream detector (YOLOv11 [19]). This framework explicitly separates illumination enhancement and structural edge preservation, addressing the limitations of prior methods. The adaptive illumination enhancement (AIE) module enhances local contrast and adjusts global brightness distributions through contrast-limited adaptive histogram equalization (CLAHE) and gamma correction, while the EdgeSeal enhancement (ESE) module reinforces structural edges and defect boundaries using morphological operations and sharpening filters.
Although the individual operations employed in AIE and ESE are well-established in the literature, the originality of this study lies in how these operations are modularly organized, adaptively tuned, and seamlessly integrated into the detection pipeline. Unlike conventional approaches that apply enhancement operators in a fixed sequence, the proposed framework optimizes the cooperation between illumination correction and edge reinforcement to ensure stability across diverse low-light environments.
This modular design offers a practical and generalizable preprocessing solution that can be embedded into existing object detection systems without architectural modification or retraining. The proposed framework is integrated with the YOLOv11 detection model and evaluated on a rail surface defect dataset, achieving an ~7% increase in mean average precision (mAP) across three defect categories (defects, dirt, and gaps) compared with the baseline model. The main novelty of this work, therefore, lies not in the use of any specific enhancement algorithm but in the design of a flexible and adaptive preprocessing architecture that significantly improves the robustness of YOLO-based rail defect detection under low-light and degraded-visibility conditions. The key contributions of this paper are as follows.
  • AIE using CLAHE and gamma correction effectively improves local contrast and adjusts global brightness distributions under low-light conditions.
  • ESE, based on a morphological edge-focused enhancement strategy, preserves fine defect contours and sharpens structural boundaries to support high-fidelity feature extraction.
  • When integrated with YOLOv11, the proposed preprocessing framework leverages its modular and adaptive design to enhance defect detection performance across diverse low-light and degraded-visibility conditions, achieving up to 7% increases in mAP over the baseline model.

2. Related Works

2.1. Vision-Based Rail Defect Detection

Deep learning has markedly advanced automated rail surface inspection, facilitating the development of lightweight and efficient models suitable for real-time deployment. For instance, GhostMicroNet-YOLOv8n was introduced for fast and energy-efficient damage detection on rails [20]. Similarly, improved YOLOv8 architectures, optimized for precision and speed, have achieved detection accuracies of up to 94.9% [21].
Attention mechanisms and feature fusion strategies have also been employed to enhance defect localization, e.g., a CenterNet-based approach with a ResNeXt backbone was used to extract multiscale features, enabling the detection of small defects [22]. Other works, such as those using DP-YOLO, incorporated depthwise separable convolutions and attention modules to reduce computational load while maintaining high precision in fastener defect detection [23,24,25]. Multimodal learning has emerged as a promising direction, with models such as FusWay combining vision and audio signals using vision transformers and YOLOv8 to improve robustness in real-world rail inspection scenarios [26]. Although these methods demonstrate competitive performance under ideal lighting conditions, their accuracy tends to degrade in low-illumination environments because of reduced contrast, weak texture visibility, and loss of fine edge structures.

2.2. Low-Light Image Enhancement for Industrial Applications

The enhancement of image quality under poor lighting has been explored across various industrial domains to improve detection accuracy. Retinex-based approaches, such as KinD [27] and Retinexformer [28], decompose images into reflectance and illumination components to achieve better restoration. GANs have been employed to synthesize clearer and brighter images. For example, CES-GAN effectively restored low-light fastener images using a U-Net architecture with adversarial training [29]. BrightsightNet [30] proposed a lightweight U-Net–based network designed for low-light railway image enhancement, which effectively learns both structural and illumination and structural cues to support reliable defect detection without explicit preprocessing. Similarly, LFGB-YOLO [31] integrates a low-light enhancement module directly within the detection architecture, achieving improved precision and recall for rail fastener detection under challenging illumination conditions. Moreover, Enhanced YOLOv5 [32] combines illumination correction and defect detection within a unified framework, ensuring stable crack detection performance in tunnel environments with insufficient lighting. Other studies have focused on anomaly detection under weak lighting conditions, highlighting the importance of preserving texture fidelity for reliable detection [33]. However, most of these methods focus solely on image enhancement without integration into detection pipelines or suffer from oversmoothing, which results in the loss of critical boundary details.

2.3. Research Gap and Our Approach

Despite the notable advancements in the field of vision-based rail defect detection, several limitations remain.
  • Global enhancement techniques often fail to preserve fine textures and boundary features essential for detecting microdefects.
  • Most detection models are not optimized for illumination variability, which limits their effectiveness in real-world low-light scenarios.
  • Few works combine local contrast enhancement and structural edge reinforcement within a unified preprocessing framework.
To address these limitations, we developed a two-stage preprocessing approach integrating AIE and ESE modules to simultaneously enhance image brightness and preserve defect boundary integrity. This framework, when integrated with a state-of-the-art detection model, delivered substantial accuracy improvements across various defect types and challenging environmental conditions.

3. Methodology

Low-light images of rail surfaces often suffer from degraded quality, which hinders the accurate identification of rail defects. Images captured under low-illumination conditions, e.g., at night, inside tunnels, or in backlit environments, typically exhibit low brightness, poor contrast, and blurred boundaries of rail defects, which markedly reduces detection accuracy [34]. The various image enhancement techniques used to address this issue often fail to preserve the visual information pertaining to small and fine-grained defects. Specifically, simple contrast adjustments or global filtering techniques are insufficient to retain the structural characteristics and contours of rail defects [35].
Herein, we propose a modular image preprocessing framework that combines AIE and ESE to effectively preserve and highlight visual information pertaining to rail defects, even under low-light conditions, primarily aiming to maximize the detection performance of the YOLOv11 object detection model based on enhanced image quality [36]. The proposed preprocessing-based rail defect detection framework features three main stages. AIE enhances brightness and contrast to restore the visual information pertaining to dark regions within the rail images. ESE emphasizes the contours and boundaries of rail defects, increasing the prominence of their structural features. Finally, the preprocessed images are input into the YOLOv11 object detection model to precisely detect the location and type of rail defects. The architecture of the proposed framework is shown in Figure 1, and the details of each module are described in the following sections.

3.1. Adaptive Illumination Enhancement

Low-light rail images often suffer from low overall brightness and insufficient contrast between defects and their surrounding background, which hinders visual defect identification. As a result, defect shapes and boundaries become blurred, and the performance of subsequent processes, such as binarization, defect detection, and object recognition, deteriorates. Therefore, an effective preprocessing step that compensates for the brightness and contrast characteristics of input images is crucial for enhancing rail defect detection accuracy. To address these challenges, we developed an AIE module that sequentially applies CLAHE and gamma correction. Although CLAHE and gamma correction are established image enhancement techniques, their adaptive integration within the AIE module, in coordination with the ESE module, allows the framework to optimally balance local contrast improvement and global brightness adjustment, ensuring robustness across diverse low-light environments.

3.1.1. CLAHE

CLAHE works by dividing the input image into S nonoverlapping subregions of size m × m . The pixel intensity distribution within each region is independently analyzed and adjusted to improve local contrast. For each region i , the histogram value H i k is normalized as H i n k = H i k / N i , where N i is the total number of pixels in region i . A cumulative distribution function (CDF) is then calculated for each region, with the cumulative probability for intensity level k given by
C D F i k = j = 0 k H i n ( j ) .
To avoid noise amplification, a clipping limit ( C l i m ) is applied, yielding a clipped histogram H i c k = min H i n k , C l i m . The CDF of the clipped histogram is then recalculated as
C D F i c k = j = 0 k H i c j N i .
Finally, the pixel intensity is remapped using the minimum CDF value, C D F i , m i n , with the result given by
I c l a = C D F i c k C D F i , m i n 1 C D F i , m i n × L 1 ,
where L is the total number of possible pixel intensity levels (typically 256). This process ensures that local contrast is enhanced uniformly across each region, making defects more visually distinct in low-light rail images.

3.1.2. Gamma Correction

Gamma correction adjusts global brightness, enhancing visual information in low-light areas. Although CLAHE improves local contrast and emphasizes fine defect details, it may struggle with shadowed regions or subtle cracks, where boundaries are not clearly defined. To address this problem, gamma correction was applied after CLAHE processing to further amplify the brightness of dark regions and thus enhance defect visibility.
The image I c l a processed by CLAHE was first normalized to a range of 0–255. Then, the gamma correction was applied as
I g m a = 255 × ( I c l a 255 ) γ ,
where γ is the gamma coefficient that is typically set to γ < 1 to enhance the brightness of dark pixels. The combined effect of CLAHE-based local contrast enhancement and gamma correction–based global brightness adjustment improved the overall contrast of rail images, increasing defect detection accuracy. Figure 2 compares the original, CLAHE-processed, and post-gamma-correction (final) images.

3.2. EdgeSeal Enhancement

Despite the improvement in brightness and local contrast achieved using the AIE module, the structural form and boundaries of fine defects in low-light rail images remained unclear. Cracks, wear, and damage in rail rails are often characterized by small size, low intensity, and weak contrast with the background, which hinders visual identification based on simple illumination correction techniques alone. As a result, the incidence rates of false positives (FPs) and false negatives (FNs) in defect detection increase.
To address this problem, we proposed an ESE module combining morphological closing and sharpening operations. This module was designed to preserve the structural boundaries of defects while suppressing background noise, providing clearer edge information during the feature extraction stage of object detection.

3.2.1. Morphological Closing

Morphological operations are based on the geometric structure of objects within an image and used to correct their boundaries and structural continuity. Among these, the closing operation is particularly useful for enhancing defect visibility in low-light conditions, sequentially applying dilation and erosion operations to fill small holes within an object and smooth its outer contour. This operation is mathematically defined as
I m o p = I g m a B B ,
where B is a structural kernel; denotes the dilation operation, where the pixel is set to unity if any part of the image overlaps with the kernel; and represents the erosion operation. The dilation operation expands the boundary of the object and is effective in filling irregular defect contours or small holes within the defect region. The erosion operation contracts the object boundary, effectively restoring the original shape while removing unnecessary weak noise or isolated pixels.
Through this sequence, the closing operation flattened low-frequency image components, enhancing the structural continuity of defects while suppressing background noise. This process was especially effective in revealing the contours of fine defects, such as microcracks on rails, which were challenging to detect in low-light images.

3.2.2. Sharpening

Although morphological closing successfully corrected the overall contours of rail defects, finer details such as boundary lines, textures, and structural information remained blurred, which is critical for detecting small defects like micro-cracks. To further enhance the clarity of these high-frequency features, sharpening was applied after morphological processing.
Sharpening emphasized the high-frequency components of images, such as boundaries and textures, thereby making the defect contours sharper and more distinct from the background. The first stage involved applying Gaussian blurring is represented as
I b l u = I m o p * G σ ,
where G σ is a Gaussian kernel with standard deviation σ used to blur the image and “ * ” denotes the convolution operation, which applies the kernel to the image I m o p . The high-frequency component I h i was extracted by I m o p I b l u , which contained the edges, texture, and finer structural details of the defects and captured the essential information pertaining to defect contours and small structures critical for detecting microdefects in rails. Finally, I h i was combined with I m o p to produce the sharpened image I s h p as
I s h p = I m o p + α · I h i ,
where α is a scaling factor that controls the intensity of the high-frequency component in the final sharpened image and thus helps adjust the degree to which defect details are emphasized. The visual effects of these enhancements are shown in Figure 3, which illustrates the improvement in defect visibility and enhanced sharpness of defect boundaries, validating the effectiveness of the proposed approach.

3.3. Object Detection

Object detection simultaneously estimates the class and location of objects within an image by predicting the bounding box and class information for each object. Traditional object detection techniques, which feature two separate stages, have been superseded by end-to-end integrated approaches performing both tasks simultaneously. Among these, the YOLO series is a prominent object detection model, dividing the input image into a grid and predicting the presence of an object, its class, and the bounding box coordinates within each grid cell. This architecture offers the advantages of high processing speed and accuracy and is therefore suitable for real-time applications. Furthermore, numerous structural improvements aimed at reducing model size while enhancing accuracy have been proposed [37].
Herein, we adopted the latest version of the YOLO series, YOLOv11, as the rail defect detection model. YOLOv11 is designed to achieve model lightweighting and high-precision inference performance, which is particularly beneficial for the detection of small defects and unclear boundaries, such as those in rail images captured under low-light conditions. The overall network architecture of YOLOv11 features three modules, namely the backbone, neck, and head (Figure 4).
The backbone is the module responsible for extracting features from the input image at various levels and employs several lightweight blocks that balance computational efficiency with representational power. Key components include the traditional convolution block (CBS: convolution + batch normalization + SiLU), as well as more specialized modules such as C3k2, C2PSA, and spatial pyramid pooling fast, which enhance feature extraction across multiple scales. These components allow the model to capture meaningful features at different resolutions, thereby providing a foundation for detecting defects of varying sizes.
The subsequent neck module integrates high- and low-level features to effectively generate multiscale feature maps and combines the feature pyramid network and path aggregation network to optimize feature fusion. Finally, the head module outputs the detection results, predicting the object class, bounding box coordinates, and object presence probability for each grid cell. This process is trained based on a multiloss function given by
L t o t = λ c l s · L c l s + λ b o x · L b o x + λ o b j · L o b j ,
where L c l s , L b o x , and L o b j represent the classification loss, bounding box regression loss, and objectness loss, respectively. L c l s evaluates how well the predicted object class matches the actual label, L b o x quantifies the alignment between the predicted bounding box and true object position and size, and L o b j determines the presence of an object at the predicted location. λ c l s , λ b o x , and λ o b j are weights controlling the importance of each loss term. To improve the precision of bounding box localization, the complete intersection over union loss method was employed. Unlike the traditional intersection over union, this method considers not only the overlap between boxes but also the distance between their center points and their aspect ratios, enabling more accurate bounding box regression.

4. Experimental Results

To conduct the experiments, we utilized the Rail Defect dataset [38], which includes low-light images of real rails. The original and preprocessed images were tested using the same YOLOv11 model, and its defect detection performance was assessed in terms of precision, recall, and mAP. By analyzing the performance differences between various comparison models, we aimed to validate the effectiveness of the proposed method.

4.1. Experimental Setup

Experiments were conducted using a set of real rail surface images acquired in low-light environments. The dataset included various defect types, primarily structural damages and surface contaminations. Initially, eight flaw classes were annotated, namely dents, crushes, scratches, slants, damages, unknown, dirt, and gaps. Owing to high visual similarity and inherent class imbalance among six of these classes (dents, crushes, scratches, slants, damages, unknown), they were consolidated into a single class labeled “defects”. Consequently, the final classification was performed over three target classes, namely defects, dirt, and gaps. Representative samples for each class are shown in Figure 5.
The dataset (440 images), collected from operational railway environments including tunnels and nighttime scenes with low-light and uneven illumination, was divided into training (318), validation (42), and testing (80) subsets. Three defect classes were annotated for evaluation: Defect (200), Dirt (162), and Gap (78). All images were resized to 640 × 640 pixels to ensure consistency in input dimensions. The CLAHE operation was applied with a clip limit of 2.0 and a tile grid size of 8 × 8 to enhance local contrast while avoiding over-amplification of noise. Gamma correction was uniformly applied with a γ value of 1.2, effectively brightening underexposed regions. Morphological operations for edge preservation employed a 3 × 3 kernel for both erosion and dilation. Gaussian blur was implemented with a 5 × 5 kernel and a standard deviation of 1.0 to suppress high-frequency noise, followed by a sharpening factor of 1.5 to restore structural sharpness and contour integrity.
Experiments were conducted on a machine running Windows 10 and featuring two Intel Xeon Silver 4310 processors, 16 GB of DDR3 memory, and an NVIDIA GeForce RTX 1080 graphics processing unit. Model training was conducted using the Adam optimizer with a learning rate of 1 × 10−4 and weight decay of 1 × 10−5. Each model was trained for 100 epochs with a batch size of 16, and early stopping was applied with a patience of 10 epochs based on the validation mAP. To improve model generalization, data augmentation techniques including horizontal and vertical flips, random rotations of ±15°, and brightness adjustments of ±10% were applied during training.
Recall is defined as the number of correctly detected defects divided by that of all defects. This metric reflects the ability of the model to minimize detection omissions (FNs) and can also be expressed as recall = true positives (TPs, i.e., number of correctly detected defects)/(TPs + FNs). Precision is defined as the number of true defects divided by that of model-detected defects. This metric quantifies the ability of the model to minimize false detections (FPs) and can also be expressed as precision = TPs/(TPs + FPs), with higher values corresponding to lower FP rates.

4.2. Ablation Study

To quantitatively evaluate the individual and combined contributions of the proposed AIE and ESE modules, we conducted a series of ablation experiments. Table 1 summarizes the detection performance under different module configurations.
Under the baseline condition without either module, the Precision for Defect, Dirt, and Gap classes was 0.84, 0.85, and 0.98, respectively, with corresponding Recall values of 0.79, 0.61, and 1.00. The overall mAP@50 and mAP@50–95 were measured as 0.85 and 0.55, respectively.
When the AIE module was applied alone, the Precision for the Defect class increased to 0.88, while its Recall decreased to 0.65, reflecting a slight compromise in structural detail recognition. The Dirt class exhibited a small decrease in Precision (0.85 → 0.82) but a significant increase in Recall (0.61 → 0.78), while the Gap class remained similar to the baseline. Overall, the mAP@50 increased to 0.87, indicating that the AIE effectively mitigates local brightness inconsistencies and enhances contrast under low-light conditions. However, excessive enhancement may lead to texture distortion, particularly affecting fine structural anomalies.
When the ESE module was applied alone, Precision values for the Defect and Dirt classes increased to 0.90 and 0.91, respectively, while Recall remained similar to the AIE-only configuration (0.66 and 0.78). The overall mAP@50 and mAP@50–95 were 0.88 and 0.56, respectively, demonstrating comparable performance gains. These results indicate that ESE effectively sharpens structural edges and preserves object contours, suppressing false positives and improving boundary recognition. Nevertheless, over-amplification of edges may slightly distort fine boundary textures, potentially reducing structural detail recognition for specific classes.
Integrating both AIE and ESE modules led to consistent performance improvements across all defect categories. Precision and Recall increased for the Defect and Dirt classes, while the Gap class achieved near-perfect scores. The overall mAP@50 and mAP@50–95 reached 0.92 and 0.59, respectively, highlighting a clear improvement over the baseline. These results indicate that AIE and ESE complement each other: AIE enhances visual cues by adjusting local contrast, while ESE reinforces structural contours and mitigates noise amplification. The combined effect effectively balances the Precision–Recall trade-off, enabling robust and accurate detection under challenging low-light conditions.
Table 2 presents repeated-run evaluations of the baseline YOLOv11 and the proposed AIE + ESE-based YOLOv11. The proposed framework achieved an average mAP@50 of 92.6 ± 0.2, significantly outperforming the baseline (85.4 ± 0.2). The low standard deviation confirms that the proposed enhancement framework delivers stable and reliable performance across multiple trials.
Collectively, these quantitative findings demonstrate that the proposed preprocessing modules substantially improve defect visibility and detection performance under low-light conditions. The complementary effects of AIE and ESE preserve structural integrity, enhance local contrast, and suppress irrelevant features, thereby supporting robust and generalizable detection across diverse defect types.
In addition, to clearly determine the relative advantages of the proposed AIE and ESE modules over existing preprocessing techniques, additional comparative experiments were conducted. For this purpose, the YOLOv11 baseline model was evaluated using datasets preprocessed by traditional image enhancement methods—CLAHE, histogram equalization (HE), and dark channel prior (DCP)—as well as the recent deep learning–based enhancement method LEGAN [39]. Table 3 presents the comparative detection performance results under general low-light conditions.
Traditional approaches such as HE, CLAHE, and DCP provided moderate improvements in brightness and contrast but often led to structural distortion or uneven enhancement, resulting in fluctuating precision–recall balance. The deep learning–based LEGAN method restored brightness and texture more effectively but introduced artificial details, which increased false detections in the defect class. In contrast, the proposed modular preprocessing framework, integrating AIE and ESE, achieved the most stable and superior results (mAP@50 = 0.92, mAP@50–95 = 0.59). This improvement originates from the complementary mechanisms of the two modules: AIE effectively mitigates illumination imbalance and enhances local contrast, while ESE reinforces structural edges and preserves fine defect contours. Consequently, the proposed preprocessing pipeline substantially improves the visual quality of low-light images and enables more reliable and balanced defect detection, regardless of the downstream detection model.

4.3. Comparative Evaluation

To validate the performance of the proposed image preprocessing–based detection framework, we conducted comparative experiments using several state-of-the-art object detection models, namely YOLOX [40], RT-DETR [41], YOLOv11 [19], improved YOLOX [42], and SMI-YOLOv8 [43]. This selection was made to provide a comprehensive comparison: baseline detection models (YOLOX, RT-DETR, YOLOv11) illustrate the limitations of standard architectures, improved YOLOX and SMI-YOLOv8 represent prior attempts to address low-light detection challenges, and YOLOv11 serves as our baseline for evaluating the benefit of combining it with the proposed preprocessing modules. This arrangement clarifies the goal of the experiment—to reveal the inherent limitations of detection-only models and to prove that traditional object detection models, when used without preprocessing, struggle to detect subtle and low-contrast rail surface defects. All models were evaluated under identical experimental conditions using the metrics and defect categories described above (Table 4). Specifically, all models were trained on the same dataset with the input image resolution of 640 × 640, the batch size of 16, 300 training epochs using the SGD optimizer with a momentum of 0.973 and a weight decay of 0.005, ensuring a controlled and reproducible evaluation.
The YOLOX series models [40,42] were generally inferior to recent architectures. Specifically, the accuracy of defect class detection was notably low, primarily because of the difficulties in recognizing fine-grained structural anomalies. Additionally, both YOLOX variants showed limited recall for the gap class, which indicated the frequent omission of TPs. Although high precision was achieved for the dirt class, the corresponding recall was low, suggesting a substantial number of missed TPs. Consequently, the mAP@50 scores for YOLOX and improved YOLOX remained around 0.5, underscoring the limited effectiveness of these models in low-light environments.
RT-DETR exhibited excellent detection capabilities for larger or well-defined defects, achieving high recall in the gap category. However, its performance on the dirt class was suboptimal, probably because the visual similarities between dirt patterns and the background decreased detection reliability. YOLOv11 achieved stable performance on the defect and gap classes, but its recall on the dirt class was lower, resulting in an overall mAP performance similar to that of RT-DETR.
In contrast, the proposed model outperformed all baseline models across the three defect categories. Specifically, it achieved a recall of 0.90 for the defect class, indicating a high sensitivity to subtle and fine-grained structural anomalies. For the dirt class, both precision and recall were notably improved, reaching 0.92 and 0.89, respectively. Furthermore, our model attained perfect recall for the gap class. These results contributed to the highest overall mAP@50 score of 0.92 and were attributed to the synergistic effect of the proposed preprocessing pipeline. The AIE module effectively corrected local brightness inconsistencies and enhanced contrast in underexposed regions, thereby improving the visibility of defects in low-light conditions, while the ESE module preserved and sharpened defect contours, helping the detection model learn more discriminative features. Collectively, these enhancements mitigated the limitations imposed by challenging lighting conditions and substantially improved the robustness and generalization ability of the model across various defect types.
Figure 6, Figure 7 and Figure 8 present qualitative detection results for defect, dirt, and gap classes, respectively, with blue bounding boxes representing ground truth (GT) annotations, green boxes indicating correct detections (TPs), and red boxes denoting misdetections (FPs). Figure 6 shows that the YOLOX model frequently failed to detect fine-grained defects or generated multiple overlapping bounding boxes for a single defect instance, indicating poor localization capability. Other baseline models showed improved localization capabilities but lacked precision in delineating defect boundaries. In contrast, the proposed model exhibited superior boundary adherence and spatial accuracy, successfully detecting subtle defects with minimal FPs. Figure 7 shows that YOLOX and improved YOLOX detected only a limited subset of dirt instances, thus featuring high FN rates. YOLOv11 showed improved recall but failed to detect certain dirt regions. Although RT-DETR successfully identified all dirt regions, it also produced several FPs in nondefective areas, probably because of background similarities. The proposed model accurately detected all dirt instances without FPs, demonstrating high precision and robust discriminative capability. Figure 8 shows that YOLOX entirely failed to detect gap defects, while improved YOLOX generated several FPs on nondefective areas. In contrast, RT-DETR, YOLOv11, and the proposed model accurately detected gaps. However, the proposed model demonstrated the most precise alignment between predicted bounding boxes and the GT with no false detections and accurate localization.
Collectively, these qualitative results reveal that the proposed preprocessing framework substantially improved defect visibility under low-light conditions, enabling more accurate and robust detection across all defect categories. The ability of our model to preserve structural integrity and suppress irrelevant features is particularly effective in visually challenging environments.
To better reflect real-world railway operating conditions, we extended the simulation beyond low-light environments to include additional visibility degradation factors such as fog and shadows, which commonly occur in tunnels, during early morning hours, or under rainy conditions and often obscure the visual features of defects and hinder accurate detection. Consequently, evaluating model performance solely under low-light conditions is insufficient to ensure practical applicability. Hence, further experiments were conducted using the object detection models listed in Table 3 for test images modified to simulate fog and shadow conditions (Table 5). The proposed model maintained its high detection performance across various visual disturbances, which verified its robustness and generalization capability in real-world operating environments.
The baseline models, including the YOLOX family [40,42], RT-DETR [41], and YOLOv11 [19], exhibited notable precision and recall degradation across defect and dirt classes under fog conditions. Notably, YOLOX [40] demonstrated a severe decline in defect recall, which decreased to 0.01 and indicated an almost complete inability to detect defects in fog-affected images. In contrast, the proposed model maintained a substantially higher performance, achieving precision and recall values of 0.95 and 0.64, respectively, for the defect class and 0.97 and 0.73, respectively, for the dirt class. For the gap class, the proposed model preserved near-optimal performance, achieving precision and recall values (0.95 and 1.00, respectively) comparable with those observed under clear conditions.
Overall, the proposed model attained an mAP@50 of 0.87 and mAP@50–95 of 0.52, exhibiting a slight performance decline relative to clear conditions but still substantially outperforming all baseline models. These findings demonstrate that the integration of the AIE and ESE modules effectively preserved and enhanced the visual characteristics of defects even in visually degraded environments such as fog and confirm the robustness and generalization capability of the proposed approach in handling diverse real-world visibility challenges frequently encountered in railway operations.
Figure 9 illustrates qualitative detection results for each model under fog-augmented low-light conditions, corresponding to the comparative analysis of the defect class detailed in Table 4. All baseline models exhibited a notable number of FPs and FNs, particularly for subtle or small-scale defects. These results highlight the sensitivity of conventional object detection models to visibility degradation. In contrast, the proposed model delivered robust and accurate detection, maintaining clear localization and minimizing FPs and FNs, which confirmed its effectiveness in adverse visual environments.

5. Discussion

The preprocessing framework designed to enhance the efficiency of rail surface defect detection under low-light conditions integrated AIE and ESE modules to address the common challenges of poor visibility, low contrast, and blurred defect boundaries in real-world railway environments. When integrated with YOLOv11, this framework achieved a notable detection performance improvement, as revealed by an ~7% increase in mAP@50 compared with the baseline model.
This performance gain was primarily attributed to the synergistic effect of the preprocessing steps: CLAHE-based local contrast enhancement improved the visibility of faint defect regions; gamma correction effectively adjusted global brightness distributions, particularly benefiting dark areas; and morphological sharpening preserved edge integrity and enhanced structural features. The proposed model maintained high precision and recall across all tested defect classes, which confirmed its robustness and practical potential for deployment in real rail inspection systems. In comparative evaluations, our model consistently outperformed other advanced models, including YOLOX variants and transformer-based detectors such as RT-DETR. Moreover, the model retained its performance advantage under degraded-visibility conditions with simulated fog and shadow effects, which further validated its generalization capability in challenging real-world scenarios.
A notable drawback of our preprocessing technique is that it enhances all visual features indiscriminately, including both defect regions and nondefect elements such as gravel, bolts, or debris surrounding the tracks. This feature can lead to increased FPs, as nondefective regions with similar edge or texture patterns can be mistakenly classified as defects. This trade-off highlights the need for more selective feature enhancement methods in future framework iterations. Additionally, the evaluation was conducted on a relatively small-scale dataset comprising 440 images. This limited dataset may not fully reflect the diversity of real-world railway conditions, which can widely vary in terms of lighting, weather, background noise, and defect types. To establish broader applicability and improve robustness, the framework should be validated on larger and more diverse datasets representing various operating environments. A further limitation lies in the image-dependent nature of the enhancement procedure, which may introduce performance variability under extreme lighting conditions. Although the proposed enhancement modules operate based on the characteristics of input illumination, their parameters were empirically optimized to ensure consistent performance under various lighting conditions. The modular structure also allows straightforward adjustment of enhancement parameters for new datasets or environments, mitigating potential variability and enhancing the framework’s generalizability.
Based on the abovementioned findings and limitations, the following future research directions are proposed. First, segmented learning strategies or class-specific enhancement techniques should be introduced to better learn the visual feature differences between background and defect objects. Second, the preprocessing technique should be extended from a static to a dynamic learning–based approach adapting according to the conditions to improve adaptability to varying illumination conditions. Third, beyond rail defect detection, one should explore the applicability of our framework to other industrial fields, such as road crack detection, industrial pipeline leak monitoring, and aircraft surface inspection.
In conclusion, although the proposed framework offers notable improvements in low-light rail defect detection, continued refinement and expansion are required to fully realize its potential for real-world deployment in safety-critical infrastructure monitoring systems.

6. Conclusions

A preprocessing-based detection framework is constructed to enhance defect detection performance in real-world railway environments affected by poor illumination. This framework consists of AIE and ESE modules, which improve the visual quality of images through local contrast enhancement and nonlinear brightness correction and markedly promote the recognition of fine defects by preserving defect contours and increasing structural sharpness. The proposed approach outperforms state-of-the-art object detection models, including YOLOv11, achieving competitive results with an mAP@50 of 92% and mAP@50–95 of 59%. Importantly, this method enhances detection performance without requiring modifications to the underlying object detection architecture and is therefore broadly applicable to various models and use cases.
The proposed framework enables reliable and accurate rail defect detection under challenging lighting conditions and holds promise as a core technology for the real-time monitoring, safety management, and automated maintenance of railway infrastructure. Future works should focus on developing more refined preprocessing strategies considering various illumination and environmental conditions, as well as expanding the framework to other industrial fields.

Author Contributions

Conceptualization, G.B.; Methodology, G.B.; Software, G.B.; Validation, S.Y.; Formal analysis, G.B. and S.Y.; Investigation, S.Y. and J.C.; Data curation, G.B.; Writing—original draft, G.B.; Writing—review and editing, S.Y. and J.C.; Supervision, J.C.; Project administration, J.C.; Funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by a National Research Foundation of Korea (NRF) grant funded by the Korean government (MOE) (No. 2021R1I1A3055973) and the Soonchunhyang University Research Fund.

Data Availability Statement

The data presented in this study are available in https://github.com/sirlis/RailDefect (accessed on 29 November 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Silva, P.; Ribeiro, D.; Pratas, P.; Mendes, J.; Seabra, E. Indirect assessment of railway infrastructure anomalies based on passenger comfort criteria. Appl. Sci. 2023, 13, 6150. [Google Scholar] [CrossRef]
  2. Zhu, W.; Han, X.; Zhang, K.; Lin, S.; Jin, J. Application of YOLO11 model with spatial pyramid dilation convolution (SPD-Conv) and effective squeeze-excitation (EffectiveSE) fusion in rail track defect detection. Sensors 2025, 25, 2371. [Google Scholar] [CrossRef]
  3. Xiong, L.; Jing, G.; Wang, J.; Liu, X.; Zhang, Y. Detection of rail defects using NDT methods. Sensors 2023, 23, 4627. [Google Scholar] [CrossRef]
  4. Miao, H.; Zhu, Q.; Zhang, Y.; Kan, Q.; Wang, P.; Zhang, W.; Kang, G. SH Guided wave excitation in rails for defect and stress monitoring. Mech. Syst. Signal Process. 2025, 224, 112064. [Google Scholar] [CrossRef]
  5. Chen, Y.; Feng, B.; Kang, Y.; Cai, X.; Wang, S.; Li, Y.; Duan, Z. Automatic crack identification using a novel 3D profilometry-based magnetic particle testing method. Mech. Syst. Signal Process. 2023, 200, 110720. [Google Scholar] [CrossRef]
  6. Quercio, M.; Santoro, L.; Sesana, R.; Fulginei, F.-R. Advanced feature analysis of eddy current testing signals for rail surface defect characterization. IEEE Access 2025, 13, 141156–141169. [Google Scholar] [CrossRef]
  7. Kumar, A.; Harsha, S.-P. A systematic literature review of defect detection in railways using machine vision-based inspection methods. Int. J. Transp. Sci. Technol. 2024, 13, 100–118. [Google Scholar] [CrossRef]
  8. Choi, J.-Y.; Han, J.-M. Deep learning (Fast R-CNN)-based evaluation of rail surface defects. Appl. Sci. 2024, 14, 1874. [Google Scholar] [CrossRef]
  9. Yu, Q.; Liu, A.; Yang, X.; Diao, W. An improved lightweight deep learning model and implementation for track fastener defect detection with unmanned aerial vehicles. Electronics 2024, 13, 1781. [Google Scholar] [CrossRef]
  10. Wang, M.; Li, K.; Zhu, X.; Zhao, Y. Detection of surface defects on railway tracks based on deep learning. IEEE Access 2022, 10, 126451–126465. [Google Scholar] [CrossRef]
  11. Luo, H.; Cai, L.; Li, C. Rail surface defect detection based on an improved YOLOv5s. Appl. Sci. 2023, 13, 7330. [Google Scholar] [CrossRef]
  12. Ren, Z.; Liu, Y.; Zhao, D.; He, Y.; Zhang, J. A deep learning-driven framework for real-time recognition and quantification of subway tunnel surface defects using high-resolution imaging. Urban Rail Transit 2025, 11, 279–299. [Google Scholar]
  13. Tian, Z.; Qu, P.; Li, J.; Sun, Y.; Li, G.; Liang, Z.; Zhang, W. A survey of deep learning-based low-light image enhancement. Sensors 2023, 23, 7763. [Google Scholar] [CrossRef]
  14. Cheng, H.; Pan, K.; Lu, H.; Wang, W.; Liu, Z. Large–small-scale structure blended U-net for brightening low-light images. Sensors 2025, 25, 3382. [Google Scholar] [CrossRef]
  15. Munaf, S.; Bharathi, A.; Jayanthi, A.-N. FPGA-based low-light image enhancement using Retinex algorithm and coarse-grained reconfigurable architecture. Sci. Rep. 2024, 14, 28770. [Google Scholar] [CrossRef]
  16. Rodríguez-Rodríguez, J.A.; López-Rubio, E.; Ángel-Ruiz, J.A.; Molina-Cabello, M.A. The impact of noise and brightness on object detection methods. Sensors 2024, 24, 821. [Google Scholar] [CrossRef] [PubMed]
  17. Zhao, D.; Shao, F.; Zhang, S.; Yang, L.; Zhang, H.; Liu, S.; Liu, Q. Advanced object detection in low-light conditions: Enhancements to YOLOv7 framework. Remote Sens. 2024, 16, 4493. [Google Scholar] [CrossRef]
  18. Li, J.; Wang, X.; Chang, Q.; Wang, Y.; Chen, H. Research on low-light environment object detection algorithm based on YOLO_GD. Electronics 2024, 13, 3527. [Google Scholar] [CrossRef]
  19. Khanam, R.; Hussain, M. YOLOv11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  20. Wu, X.; Yu, S. Research on the lightweight detection method of rail internal damage based on improved YOLOv8. J. Eng. Appl. Sci. 2025, 72, 14. [Google Scholar] [CrossRef]
  21. Xu, C.; Liao, Y.; Liu, Y.; Tian, R.; Guo, T. Lightweight rail surface defect detection algorithm based on an improved YOLOv8. Measurement 2025, 232, 115739. [Google Scholar]
  22. Mao, Y.; Zheng, S.; Li, L.; Shi, R.; An, X. Research on rail surface defect detection based on improved CenterNet. Electronics 2024, 13, 3580. [Google Scholar] [CrossRef]
  23. Chen, L.; Sun, Q.; Han, Z.; Zhai, F. DP-YOLO: A lightweight real-time detection algorithm for rail fastener defects. Sensors 2025, 25, 2139. [Google Scholar] [CrossRef]
  24. Ma, S.; Li, R.; Hu, H. PGCI-YOLO: A railway fastener detection algorithm based on improved YOLOv8n. J. Real-Time Image Process. 2025, 22, 185. [Google Scholar] [CrossRef]
  25. Chen, M.; Zhang, M.; Peng, J.; Huang, J.; Li, H. A multi-category defect detection model for rail fastener based on optimized YOLOv8n. Machines 2025, 13, 511. [Google Scholar] [CrossRef]
  26. Zhukov, A.; Benois-Pineau, J.; Youssef, A.; Zemmari, A.; Mosbah, M.; Taillandier, V. FusWay: Multimodal hybrid fusion approach. Application to railway defect detection. arXiv 2025, arXiv:2509.06987. [Google Scholar]
  27. Zhang, M.; Shen, Y.; Zhong, S. SCRNet: A Retinex structure-based low-light enhancement model guided by spatial consistency. arXiv 2023, arXiv:2305.08053. [Google Scholar]
  28. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage retinex-based transformer for low-light image enhancement. arXiv 2023, arXiv:2303.06705. [Google Scholar]
  29. Cai, Y.; Liu, X.; Li, H.; Lu, F.; Gu, X.; Qin, K. Research on unsupervised low-light railway fastener image enhancement method based on contrastive learning GAN. Sensors 2024, 24, 3794. [Google Scholar] [CrossRef]
  30. Chen, Z.; Yang, J.; Yang, C. BrightsightNet: A lightweight progressive low-light image enhancement network and its application in “rainbow” maglev train. J. King Saud Univ.-Comput. Inf. Sci. 2023, 35, 101814. [Google Scholar] [CrossRef]
  31. Cai, Y.; He, M.; Chen, B. Efficient deep learning based rail fastener screw detection method for fastener screw maintenance robot under complex lighting conditions. Sci. Rep. 2024, 14, 28957. [Google Scholar] [CrossRef]
  32. Duan, S.; Zhang, M.; Qiu, S.; Xiong, J.; Zhang, H.; Li, C.; Jiang, Q.; Kou, Y. Tunnel lining crack detection model based on improved YOLOv5. Tunn. Under. Space Technol. 2024, 147, 105713. [Google Scholar] [CrossRef]
  33. He, M.; Wang, R.; Zhang, M.; Lv, F.; Wang, Y.; Zhou, F.; Bian, X. SwinLightGAN a study of low-light image enhancement algorithms using depth residuals and transformer techniques. Sci. Rep. 2025, 15, 12151. [Google Scholar] [CrossRef]
  34. Mi, Z.; Chen, R.; Zhao, S. Research on steel rail surface defects detection based on improved YOLOv4 network. Front. Neurorobot. 2023, 17, 1119896. [Google Scholar] [CrossRef]
  35. Li, G.; Han, C.; Liu, Z. No-Service Rail Surface Defect Segmentation via Normalized Attention and Dual-scale Interaction. arXiv 2023, arXiv:2306.15442. [Google Scholar] [CrossRef]
  36. Wang, T.; Kim, G.T.; Kim, M.; Jang, J. Contrast Enhancement-Based Preprocessing Process to Improve Deep Learning Object Task Performance and Results. Appl. Sci. 2023, 13, 10760. [Google Scholar] [CrossRef]
  37. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-time end-to-end object detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  38. Sirlis. RailDefect Dataset. Available online: https://github.com/sirlis/RailDefect (accessed on 29 November 2024).
  39. Tao, J.; Wang, J.; Zhang, P.; Zhang, J.; Yung, K.L.; Ip, W.H. LEGAN: A low-light image enhancement generative adversarial network for industrial internet of smart-cameras. Internet Things 2024, 25, 101054. [Google Scholar] [CrossRef]
  40. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar] [CrossRef]
  41. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs beat YOLOs on real-time object detection. arXiv 2023, arXiv:2304.08069. [Google Scholar]
  42. Zhang, C.; Xu, D.; Zhang, L.; Deng, W. Rail surface defect detection based on image enhancement and improved YOLOX. Electronics 2023, 12, 2672. [Google Scholar] [CrossRef]
  43. Dang, D.; Gong, L.; Jin, C.; Cui, X.; Yang, T.; Qin, J. SMI-YOLOv8: Intelligent detection of tunnel lining cracks via multiscale feature attention fusion. Measurement 2025, 257, 118833. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed preprocessing-based rail defect detection framework.
Figure 1. Block diagram of the proposed preprocessing-based rail defect detection framework.
Mathematics 13 03780 g001
Figure 2. Results of adaptive illumination enhancement (AIE) application: (a) input image, (b) image after contrast-limited adaptive histogram equalization (CLAHE) processing, (c) image after applying CLAHE and gamma correction.
Figure 2. Results of adaptive illumination enhancement (AIE) application: (a) input image, (b) image after contrast-limited adaptive histogram equalization (CLAHE) processing, (c) image after applying CLAHE and gamma correction.
Mathematics 13 03780 g002
Figure 3. Results of EdgeSeal enhancement application: (a) image after AIE, (b) image after morphological closing, (c) image after morphological closing and sharpening.
Figure 3. Results of EdgeSeal enhancement application: (a) image after AIE, (b) image after morphological closing, (c) image after morphological closing and sharpening.
Mathematics 13 03780 g003
Figure 4. Overall network architecture of You Only Look Once v11 (YOLOv11).
Figure 4. Overall network architecture of You Only Look Once v11 (YOLOv11).
Mathematics 13 03780 g004
Figure 5. Representative examples of annotated rail surface defects: (a) dent, (b) crush, (c) scratch, (d) slant, (e) damage, (f) unknown, (g) dirt, (h) gap.
Figure 5. Representative examples of annotated rail surface defects: (a) dent, (b) crush, (c) scratch, (d) slant, (e) damage, (f) unknown, (g) dirt, (h) gap.
Mathematics 13 03780 g005
Figure 6. Example detection results for the defect class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Figure 6. Example detection results for the defect class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Mathematics 13 03780 g006
Figure 7. Example detection results for the dirt class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Figure 7. Example detection results for the dirt class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Mathematics 13 03780 g007
Figure 8. Example detection results for the gap class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Figure 8. Example detection results for the gap class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Mathematics 13 03780 g008
Figure 9. Exemplary results of detection under low-light conditions with simulated fog for the defect class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Figure 9. Exemplary results of detection under low-light conditions with simulated fog for the defect class achieved by (a) YOLOX [40], (b) RT-DETR [41], (c) YOLOv11 [19], (d) improved YOLOX [42], (e) SMI-YOLOv8 [43], (f) the proposed model.
Mathematics 13 03780 g009
Table 1. Performance comparison of individual and integrated AIE and ESE modules.
Table 1. Performance comparison of individual and integrated AIE and ESE modules.
ModulePrecisionRecallmAP
@50
mAP
@50~95
AIEESEDefectDirtGapDefectDirtGap
--0.840.850.980.790.611.000.850.55
-0.880.820.940.650.781.000.870.56
-0.900.910.880.660.781.000.880.56
0.860.920.990.900.891.000.920.59
Table 2. Repeated run performance of the proposed and baseline models.
Table 2. Repeated run performance of the proposed and baseline models.
ModelmAP@0.50
Run 1Run 2Run 3Run 4Run 5Mean ± Std
YOLOv1185.585.885.285.385.485.4 ± 0.2
Proposed model92.992.592.692.492.892.6 ± 0.2
Table 3. Detection performance comparison of low-light image enhancement methods for rail defect detection using the YOLOv11 baseline model.
Table 3. Detection performance comparison of low-light image enhancement methods for rail defect detection using the YOLOv11 baseline model.
Image EnhancementPrecisionRecallmAP
@50
mAP
@50–95
DefectDirtGapDefectDirtGap
-0.840.850.980.790.611.000.850.55
HE0.891.000.790.620.681.000.860.57
CLAHE0.930.930.860.650.781.000.860.57
DCP0.830.930.800.730.731.000.870.56
LEGAN [39]0.790.930.880.720.741.000.860.55
Proposed AIE + ESE0.860.920.990.900.891.000.920.59
Table 4. Detection performances of our and existing object detection frameworks.
Table 4. Detection performances of our and existing object detection frameworks.
ModelPrecisionRecallmAP
@50
mAP
@50–95
DefectDirtGapDefectDirtGap
YOLOX [40]0.530.880.570.420.570.570.530.26
RT-DETR [41]0.890.770.940.730.621.000.840.51
YOLOv11 [19]0.840.850.980.790.611.000.850.55
Improved YOLOX [42]0.530.880.630.450.600.600.560.27
SMI-YOLOv8 [43]0.880.900.910.570.781.000.900.55
Proposed model0.860.920.990.900.891.000.920.59
Table 5. Detection performances of our models and existing object detection models under low-light conditions with simulated fog.
Table 5. Detection performances of our models and existing object detection models under low-light conditions with simulated fog.
ModelPrecisionRecallmAP
@50
mAP
@50–95
DefectDirtGapDefectDirtGap
YOLOX [40]0.851.001.000.010.050.420.260.05
RT-DETR [41]0.511.001.000.020.100.730.390.13
YOLOv11 [19]0.760.310.760.280.100.750.390.20
Improved YOLOX [42]0.700.170.910.420.101.000.500.32
SMI-YOLOv8 [43]0.840.871.000.590.680.730.850.46
Proposed model0.950.970.950.640.731.000.870.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bae, G.; Yoon, S.; Cho, J. Edge-Aware Illumination Enhancement for Fine-Grained Defect Detection on Railway Surfaces. Mathematics 2025, 13, 3780. https://doi.org/10.3390/math13233780

AMA Style

Bae G, Yoon S, Cho J. Edge-Aware Illumination Enhancement for Fine-Grained Defect Detection on Railway Surfaces. Mathematics. 2025; 13(23):3780. https://doi.org/10.3390/math13233780

Chicago/Turabian Style

Bae, Geuntae, Sungan Yoon, and Jeongho Cho. 2025. "Edge-Aware Illumination Enhancement for Fine-Grained Defect Detection on Railway Surfaces" Mathematics 13, no. 23: 3780. https://doi.org/10.3390/math13233780

APA Style

Bae, G., Yoon, S., & Cho, J. (2025). Edge-Aware Illumination Enhancement for Fine-Grained Defect Detection on Railway Surfaces. Mathematics, 13(23), 3780. https://doi.org/10.3390/math13233780

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop