Next Article in Journal
A Comparative Analysis of International Standards on Curved Surface Isolators for Buildings
Previous Article in Journal
Enhanced Catalytic Surfaces for Catechol Sensing: Combining Grafted Aryldiazonium Derivative with Cross-Linking Dopamine or Coupling Tyrosinase Immobilizations
Previous Article in Special Issue
Multi-Scale Feature Fusion and Context-Enhanced Spatial Sparse Convolution Single-Shot Detector for Unmanned Aerial Vehicle Image Object Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Increasing the Classification Achievement of Steel Surface Defects by Applying a Specific Deep Strategy and a New Image Processing Approach

1
Software Department, Engineering Faculty, Firat University, Elazig 23119, Turkey
2
Electric and Electronic Department, Firat University, Elazig 23119, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(8), 4255; https://doi.org/10.3390/app15084255
Submission received: 15 March 2025 / Revised: 9 April 2025 / Accepted: 10 April 2025 / Published: 11 April 2025
(This article belongs to the Special Issue Object Detection and Image Classification)

Abstract

:
Defect detection is still challenging to apply in reality because the goal of the entire classification assignment is to identify the exact type and location of every problem in an image. Since defect detection is a task that includes location and categorization, it is difficult to take both accuracy factors into account when designing related solutions. Flaw detection deployment requires a unique detection dataset that is accurately annotated. Producing steel free of flaws is crucial, particularly in large production systems. Thus, in this study, we proposed a novel deep learning-based flaw detection system with an industrial focus on automated steel surface defect identification. To create processed images from raw steel surface images, a novel method was applied. A new deep learning model called the Parallel Attention–Residual CNN (PARC) model was constructed to extract deep features concurrently by training residual structures and attention. The Iterative Neighborhood Component Analysis (INCA) technique was chosen for distinguishing features to lower the computational cost. The classification assessed the SVM method using a convincing dataset (Severstal: Steel Defect Detection). The accuracy in both the binary and multi-class classification tests was above 90%. Moreover, using the same dataset, the suggested model was contrasted with pre-existing models.

1. Introduction

Improving productivity and preserving product quality are crucial in production systems [1]. One of the most crucial aspects of quality control is the detection of surface defects. Traditional detection techniques rely on expert inspection for control, which has disadvantages such as time wastage, poor productivity, and poor reliability [2]. Specifically, automatic surface defect detection is becoming increasingly important in steel industries having extensive production fields. Due to heat treatment, non-metallic inclusion, corrosion, and emulsification, flat-rolled steel can develop flaws such as cracks, pitted surfaces, scratches, and patches [3]. Evaluating defect geometry and generating a sizable sample of statistical data is crucial for improving the surface defect process. As a consequence, automated detection and classification systems can prevent unforeseen equipment failure.
Automatic surface defect detection has frequently been accomplished using traditional machine learning and deep learning-based techniques [4]. Conventional classification approaches including model-based, statistical-based, and spectral-based methods typically rely on statistical data and visual features. These techniques, such as thresholding, Sobel, Local Binary Patterns (LBPs), Fourier transform, and Canny’s algorithm, are used to convert features obtained manually, but factors including the background, lighting, and camera angle directly impact the effectiveness of defect identification [5]. Additionally, these techniques have limitations when used on various surfaces, making them unsuitable for use in practical situations [6]. Deep learning-based methods are now used to improve defect-detection capabilities in computer vision [7]. When using deep learning techniques, the algorithm can function by producing prompt and precise predictions even in the absence of supervision. Thus, labor and time saving by detecting automatically in the steel factories have made important advantages in comparison with manual control by technicians and engineers [8].

2. Related Studies

The number of studies using machine learning to identify surface defects of steel has increased during the past few years. Martins et al. [9] performed automatic surface fault detection of rolled steel using artificial neural networks with image processing. The classification of certain defects, including clamps, holes, and welding, was detected by the image analysis Hough Transform method, while other complicated defects, such as exfoliation, oxidation, and waveform, were detected by applying Self-Organizing Maps and Principal Component Analysis to extract features. The system achieved an overall accuracy of 87% after managing real-world datasets. Li et al. [10] designed a feature fusion-based method to improve steel surface defect detection in their proposed model. In this method, a multiscale feature extraction (MSFE) strategy is adopted from a YOLO-based model. With the MSFE algorithm, features with different scales are extracted from multidimensional kernels of different convolution layers. An efficient feature fusion technique is then used to maximize feature discriminability. This model, developed on the publicly available NEU-DET dataset, achieved an optimum accuracy of 73.08%. Pang and Tan [11] developed a graph neural network-based method for detecting steel surface defects. They also used a novel attention mechanism called HDmA in this approach. This strategy was also successful in detecting defects in different fields of view. This method was tested on NEU-DET and GC-10 datasets and achieved 79.04% and 66.93% accuracy, respectively. Zhang et al. [12] presented a model based on YOLO v5 for detecting defects on steel surfaces. A multi-feature fusion technique, Res2Attention blocks, was used to improve the performance of the model. Model performance was tested on the NEU-DET and GC10-DET datasets. The classification accuracies were 78.5% and 67.3%, respectively. A vision-based automatic detection method with three-section defect detection, region extraction, and industrial liquid quantification was proposed by Zhao et al. [13]. They discovered that industrial liquids were measured with an accuracy rate of 90%. The accuracy rate of the recognition of cracks and scratches was obtained as 91% or more. By creating a dataset with six defects: scars, scratches, inclusions, burrs, seams, and iron scales. Li et al. [14] explored the surface defect recognition of steel strips by converting You Only Look Once (YOLO) completely into convolutional layers. The rates of detection, recall, and mAP were 99%, 95.86%, and 97.55%, respectively. Fu et al. [15] applied the SqueezeNet model by pre-training on the dataset of ImageNet for the classification of six different defects, which are rolled-in scales, inclusions, patches, scratches, crazing, and pitted surfaces. They came to the conclusion that the learned features significantly outperformed the hand-crafted ones for datasets that were heterogeneous and unseen. By using the strong time-sequenced properties of defects, Liu et al. [16] studied steel defect detection periodically by using a convolutional neural network (CNN) and a long short-term memory (LSTM) to detect roll mark defects. After extracting the defect features of samples with CNN, vectors were added to the long- and short-term memory to detect defects. As a result, the proposed method reached 81.9% of the performance of defect detection and outperformed the CNN approach. The performance of the system was raised to a rate of 6.2% by enhancing the attention mechanism. Liu et al. [17] utilized the dataset of NEU-CLS to improve the concurrent convolutional neural network (ConCNN) with light weight by using different scales of samples. The method’s accuracy performance was found to be 98.89% with a duration of 5.58 ms. The classification pre-training of surface defects was conducted using VGG19 by Guan et al. [18]. Feature extraction from the various levels of the defect weight model was performed using VGG19. Following that, SSIM and a Decision Tree were used to estimate the structure of VGG19 and the quality of the feature picture. The experiment used a dataset from Northeast University with six different types of steel surface defects, crazing, inclusions, patches, pitted surfaces, rolled-in scales, and scratches, each with 300 samples and a total of 1800 grayscale images. Following 23,000 steps, it was discovered that the VSD model’s validation rate was 89.86% higher than that of ResNet and VGG19. By using U-net and Deep Residual techniques for the classification of four different defects, Amin and Akhter [19] concluded that their system performances were 0.731 and 0.543 in terms of the Dice coefficient accuracy, respectively. In total, 12,568 training images and 1801 test images were created with a 1600 × 256 × 1 image size. A privacy dataset was explored by Zhao et al. [3] utilizing enhanced Faster R-CNN. By enhancing the conventional Faster R-CNN algorithm, the network structure of the Faster R-CNN was rebuilt. Following the testing, upgraded Faster R-CNN was proven to have a greater mean average accuracy performance for crazing, inclusions, patches, rolled-in scales, and pitted surface defect types with the following values of 0.501, 0.791, 0.792, 0.905, and 0.649, respectively.
After a detailed search of datasets studied at the academic level, it was understood that studies of classification are limited to surface defect detection and classification. However, surface flaws in real-world practices often affect only a small portion of the overall steel surface. As a result, using such data makes success more challenging. Because the Steel Surface Defect Database is a difficult dataset, a deep learning-based strategy that will perform well with it is used in this work.

3. Literature Gap, Motivation, and Contributions

To summarize the literature in general, in terms of deep learning, pre-trained models or lightweight CNN models have been used to classify defects on steel surfaces. Additionally, no frequency–time conversion algorithm has been used to increase the discrimination in the images. Looking at the literature, it is evident that many deep learning-based studies have been conducted to minimize defects in steel production. In studies focusing on the classification problem, the classification accuracy ranged between 80% and 90%. Especially in YOLO-based fault detection studies, the classification accuracy performance was between 65% and 75%. When these accuracy values are considered, it is obvious that the error performance of steel surfaces is still open to improvement. In a steel mill with a high production capacity, even a 2–3% increase in the defect detection performance will significantly improve the production quality. Therefore, there is a serious need for a deep learning-based application that detects steel surface defects with high performance.
This study was carried out to distinguish and classify defects on steel surfaces, as a small improvement in defect detection in high-capacity steel production will provide a large number of defect-free steel products. The proposed approach is designed in four main stages. The first stage is the processing of raw images. The second stage is responsible for extracting features from the Parallel Attention–Residual CNN (PARC) model. The third stage involves feature selection with the Iterative Neighborhood Component Analysis (INCA) algorithm. The final stage includes classification with a powerful algorithm. The proposed approach has three important contributions, as outlined below.
  • Attention and residual structures in the CNN model were added to the PARC model and trained in parallel. This increased the representation power of the features extracted from the PARC model. Therefore, the classification performance was improved. Although the parallel integration of attention and residual blocks has been studied in previous works, our proposed PARC architecture introduces a distinct combination specifically tailored for defect recognition in steel surface images. To validate its superiority, we conducted comparative experiments against baseline CNNs with only residual blocks, only attention modules, and sequential Attention–Residual designs. The results demonstrate that PARC achieves a consistently higher accuracy and robustness across both multi-class and binary classification tasks.
  • Raw images were processed with a new approach. In this approach, 1D stack data are used instead of a signal to obtain spectrogram images. For 1D stack data, the gradient of the raw images was taken and then converted into sequential 1D stack data. The gradient operation highlighted the differences in pixel values in the image. As a result, the classification performance is improved with this pre-processing procedure. Existing methods all use pixel data. In this study, thanks to the transformation of the image into 1D stack data, it was possible to access the frequency information of images that form a pattern with each other, such as scratches and cracks. In this case, it increased the classification performance compared to raw images.

4. Dataset

The suggested methodology was tested using a challenging dataset from the Kaggle database called Severstal: Steel Defect Detection (2019) [20]. Two tasks involving binary and multi-class classification used the dataset. Images of steel faults (6666 images) and images of no defects (5902 images), obtained using specialized imaging equipment to identify defects, were employed in the binary classification. The multi-class classification work used 6668 steel defect images with pitted surfaces, crazing, scratches, patches, and multi-defect class labels. Each sample of the collection was saved as a 1600 × 256 JPG image. To lessen the hardware requirements for this study, each sample was re-saved in a size of 200 × 32. Figure 1 shows a few samples for each class.

5. Proposed Methodology

The methodology of this study introduces a novel and robust approach to enhance the classification performance for surface defect detection, consisting of five distinct yet interconnected stages. Figure 2 shows the representation of the proposed approach. The innovative aspect of the approach begins with the pre-processing of raw images, where a unique method is employed. Unlike traditional image pre-processing techniques, this study applies a spectrogram algorithm to time series signals derived from images that contain surface defects. This step transforms 2D image data into a 1D stack by utilizing pixel values, enabling the extraction of spectrogram images. This novel transformation provides a powerful representation of pixel variations in defect regions, enhancing the model’s ability to capture subtle details crucial for classification.
In the second stage, the processed spectrogram images are used to train a newly designed architecture known as the Parallel Attention–Residual CNN (PARC) model. Figure 3 shows the layer connections of the PARC model. The innovation in this model lies in its dual use of attention and residual mechanisms within a customized CNN framework, operating in parallel. Attention mechanisms focus on emphasizing critical areas within the image, while residual connections allow the preservation of feature information from earlier layers by passing it forward to subsequent layers. This synergy ensures that vital details, which may have otherwise been diminished or lost in deep layers, are retained and utilized, thus improving the learning capability of the network. While state-of-the-art architectures such as Swin Transformer, EfficientNetV2, and YOLOv7 have demonstrated remarkable performance across a variety of vision tasks, they are primarily designed as general-purpose models and often involve high computational costs and large parameter counts. In contrast, the proposed PARC (Parallel Attention and Residual Convolution) model is specifically tailored for steel surface defect recognition, focusing on enhancing subtle and fine-grained texture variations that are commonly present in such industrial inspection tasks. The parallel integration of attention and residual mechanisms in PARC enables it to capture both local defect features and global structural information more effectively, without significantly increasing the computational complexity. Compared to Transformer-based models like Swin Transformer, which require large datasets and extensive training resources to generalize well, PARC achieves a competitive or superior accuracy with a lightweight and task-optimized architecture. Furthermore, unlike YOLOv7, which is object-detection-focused and might exhibit overkill or be less efficient for fine-grained classification tasks, PARC offers a better balance between precision, speed, and model complexity. Its compact design makes it more suitable for real-time deployment in edge environments commonly found in industrial inspection systems. Detailed layer information of the PARC model is given in Table A1 in the Appendix A section.
The third stage focuses on feature extraction from the fully connected (FC) layers of the PARC model, where deep features are drawn from both the activations and the processed input data. This stage sets the foundation for the next key novelty: instead of relying solely on the softmax classifier typically employed in CNN models, the extracted features are evaluated using a range of highly effective classification algorithms. This offers a fresh perspective on model training, allowing for a more versatile comparison of classifiers.
In the fourth stage, an efficient feature selection algorithm called INCA is applied. INCA stands out for its ability to significantly reduce computational costs while simultaneously improving classification performance. This step ensures that only the most relevant and influential features are retained, minimizing the overhead associated with large datasets and complex models. The Iterative Neighborhood Component Analysis (INCA) technique offers distinct advantages over more traditional feature selection and dimensionality reduction methods such as Principal Component Analysis (PCA) and Minimum Redundancy Maximum Relevance (mRMR).
Unlike PCA, which performs unsupervised dimensionality reduction by projecting data onto directions of maximum variance regardless of class labels, INCA is a supervised method that directly optimizes class separation by maximizing the classification accuracy. This makes INCA particularly effective in tasks where discriminative power is more important than variance preservation, such as defect detection or fine-grained classification.
Compared to mRMR, which focuses on selecting features that are most relevant to the target variable while minimizing redundancy between features, INCA iteratively refines the feature subset based on its impact on classification performance. This iterative refinement process allows INCA to dynamically adapt to the dataset structure and learn an optimal feature subset tailored to the classifier used, which often leads to a higher accuracy and better generalization.
In summary, while PCA and mRMR are powerful and widely used, INCA provides a more targeted and performance-driven approach to feature selection by integrating label information and classifier feedback during feature optimization.
Finally, in the fifth and final stage, this study implements seven popular classification algorithms, including Decision Tree (DT), Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Naïve Bayes (NB), Linear Discriminant (LD), Subspace KNN, Subspace Discriminate, and RUSBoosted Trees. Through extensive testing, the SVM algorithm demonstrated the best performance, highlighting the efficiency and accuracy of this method in classifying surface defects. This comprehensive methodology not only demonstrates significant advancements in pre-processing and model design but also introduces innovative techniques in feature selection and classifier evaluation, making it a substantial contribution to the field of surface defect detection and classification.

6. Images-to-Spectrogram Image Transform

Obtaining image-to-image spectrogram images for surface defect detection offers a powerful way to leverage frequency information, enhance feature representation, reduce noise, and enable the application of advanced analysis techniques, ultimately leading to more accurate and reliable defect detection.
The primary motivation for using the spectrogram lies in its ability to represent localized frequency variations over space, analogous to how time–frequency analysis is used in signal processing. Surface defects, although visually subtle, often introduce local structural irregularities that manifest as distinct frequency patterns when observed in a gradient-enhanced 1D representation. By transforming these data into a spectrogram, we can effectively capture these localized textural anomalies in a way that traditional spatial or frequency domain techniques might overlook.
Moreover, spectrograms offer a dual representation—combining gradient magnitude variations with localized frequency content—which enhances the model’s ability to distinguish fine-grained defect features. This is particularly beneficial in scenarios where defects are embedded within noisy or highly textured steel surfaces.
One of the most popular techniques in the field of signal processing is the processing of signals in the time–frequency domain. The most crucial information reveals how and when the spectral information of the signals analyzed in the time–frequency domain changes [22]. In procedures that are linearly time-invariant (LTI), like Fourier transforms, such information cannot be seen. The optimum method for observing a signal’s spectrum information in the time–frequency domain is through the use of a spectrogram. The spectrogram, which employs a sliding window to determine the Fourier transform of the signal, is often employed in the spectrum analysis of many non-stationary signals, including biological, voice, music, and seismic data [23].
The graph showing the change in 1D stack data with the gradient applied for each class is presented in Figure 4. As observed in this figure, the gradient operation yields different amplitude values depending on the type of surface defect. Particularly in the non-defect class, the gradient output values are minimal, indicating little change across the dataset for this class.
In the final step, spectrogram images were generated from the 1D stack data. The Hamming window was selected for signal windowing due to its advantages, especially its narrow main lobe and rapidly decaying side lobes in the frequency domain. These features help prevent spectral leakage, which can weaken the accuracy of spectral information. A sample size of 512 was chosen for the window, and the overlapping ratio was set at 0.125. This overlap helps reduce data loss in the 1D stack, but it also increases the computational load required to produce the spectrogram images.
To enhance the clarity of the spectral data, special attention was given to amplitude values while generating the spectrogram images. A threshold was applied to filter out points with low amplitude values, as high-amplitude points are generally more reliable in terms of spectral information. This approach simplifies the images, which is beneficial when using neural network classifiers, as less complex images tend to improve the classification performance. A 512-point FFT length was applied to ensure consistency. The resulting spectrogram images for each class are displayed in Figure 5, where distinct patterns can be observed for each class, demonstrating the effectiveness of this approach for capturing class-specific spectral features. The colormap viridis option was selected for spectrogram coloring.
Using colormap options like viridis in spectrogram visualization offers several advantages. Viridis provides perceptual accuracy, as it is designed with evenly spaced color differences that allow users to distinguish variations in data more easily. It also maintains readability in monochrome formats, ensuring that important information is preserved even when printed or viewed in low-resolution settings. The smooth, linear color transitions of Viridis avoid abrupt shifts, making it easier to interpret differences in intensity or frequency. Additionally, Viridis ensures that small changes in data are accurately represented, reducing the risk of misinterpretation, which is particularly important for scientific visualizations.
As shown in Figure 5, there is sudden color darkening or sudden appearance and disappearance of color lines in all classes except the class with no error. This shows that the Viridis option is ideal for spectrogram transformation.
The convolution layer of the CNN method intends to extract characteristic features processing input samples with convolution filters [24]. The mathematical computing of two functions is described as convolution. By conducting element-wise multiplication on each element, the convolution computing in the CNN technique implements a filter or kernel function utilized for the raw data’s transformation into processed data [25,26]. Each window’s output in a shift operation is determined by the total multiplication of the element information [27,28].
The batch normalization (BN) technique is used to design a more regular convolutional neural network. Moreover, during training, the CNN extinction gradient becomes more resistant [29,30].
In the Rectified Linear Unit layer (ReLU), the full “f(k) = max(0, k)” formula is used for all inputs as the layer activation function [31,32]. ReLU’s derivative is more appropriate and performs faster for algorithms like backpropagation as it is simpler than the sigmoid function.
The softmax function is typically used for the output in deep learning models [33]. The function converts the class scores from the fully connected layer to probabilistic values ranging from 0 and 1. The softmax function is denoted by S a j in Equations (1) and (2). It obtains an N-dimensional input vector, then produces a subsequent input vector with N-dimensional, having values between 0 and 1 [34,35]. Additionally, even though the softmax function is frequently chosen for the output layer in deep learning models, an SVM classifier can be preferred [36]. The exponential feature of the softmax function makes the differences across classes more certain.
S a j : a 1 . a n s 1 . s n  
S a j = e a j k = 1 n e a k        
Figure 6 represents the attention module utilized in this research. The gating signal vector, symbolized by g i   , has a wider scale at the feature map of output for ith layer “( x i )”, determining the focus region for each pixel [37]. Equations (3) and (4) give a computed output by applying element-wise multiplication.
o u t p u t = α i × x i
α i = σ ( φ T w x T x i + w g T g i + b g + b φ )
The linear transformations w and φ using the 1 × 1 × 1 dimensional convolution operator are the bias terms b φ and b g , respectively. The weights of the attention modules are adjusted randomly at first when the entire deep architecture is trained from end to end [38].
In conventional neural networks, each layer provides information to the next layer. Each layer enhances the subsequent layer directly for the network having residual blocks, then advances to layers, which are two to three hops away [39]. In multi-layer networks, the gradient vanishing problem is decreased by residual blocks. The two significant points of residual blocks are given below.
-
The inclusion of new layers would not harm the model performance because regularization will disregard them if not necessary.
-
When incoming layers are convenient, layer weights or kernels are not “0” because of present regularization. Therefore, the model performance can slightly improve.
The training of the attention mechanism and the residual blocks were applied concurrently in parallel in the developed PARC model. The main target is to transfer feature maps in residual and attention structures into one feature map. Hence, during the training phase, optimization determines the characteristic values from every two structures obtained from the feature map.

7. INCA Algorithm

Increased system speed without reducing approach success is the most important aim of feature selection techniques. The literature on machine learning algorithms has been enhanced by several feature selection methods [40,41,42]. Specifically, the feature selection techniques reduce the computational cost of the deep learning algorithms having a lot of features. The performance of the feature selection approach on the feature set should be profoundly assessed [43]. However, in deep learning methods, the application of this analysis method is not time-saving. While the Minimum Redundancy Maximum Relevance (mRMR) technique performs better by using a non-parametric feature set, Principal Component Analysis and the Linear Discriminant analysis techniques outperform by applying a linear feature set [44]. For classification problems, current research mostly depends on feature extraction-based algorithms in terms of feature weight relations [45,46,47].
The NCA is one of the most popular features of importance-based selection algorithms since they offer a variety of classification strategies. Additionally, the computational time for these approaches outperforms the PCA and mRMR algorithms.
Neighborhood Component Analysis (NCA), a method for feature selection and dimensionality reduction that is widely used in classification studies, is one of the most reliable supervised learning approaches for classifying multidimensional data into distinct classes [48,49]. The classification tasks performed by NCA are carried out with learning vector optimization criteria related to the categorization accuracy performance of the nearest neighbor classifier. In particular, a linear projection chosen by NCA maximizes the projected area’s performance of the nearest neighbor classifier. In NCA, training data with related class labels are applied to choose the projection that divides the classes effectively in the detected area. Nevertheless, the NCA makes assumptions about the distribution of each class, which are not reliable. It offers an equivalent fit to Gaussian mixtures for distribution modeling. To maximize the objective function F(w) for w, the regularized objective function is used in Equation (5).
F w = 1 n i = 1 n P i λ    
where the overall sample size is n, the value of the probability of ith the specimen is “Pi”, the parameter of regularization is “λ”, the dimension of the feature is “p” and the weight of the feature is “wr”. The weight values for the feature may be very near to “0” in case the selection of “λ” is performed at random. The relevant features have no importance for the method when weights are very near to zero. Thus, the parameter λ has to be arranged.
Iterative Neighborhood Component Analysis (INCA) holds advantages over Neighborhood Component Analysis (NCA) primarily due to its adaptive feature selection mechanism. Unlike NCA, INCA iteratively selects features, allowing it to dynamically adjust and optimize the learning process by choosing the most relevant features for a given task. This adaptability enhances the discriminative power, robustness, and generalization performance. By reassessing and modifying feature selections during each iteration, INCA is better equipped to handle diverse datasets and mitigate the impact of noisy or irrelevant features. The flexibility offered by INCA in tailoring its approach to specific dataset characteristics makes it a promising choice for various machine learning applications.
Choosing the parameter λ randomly may not give the best feature selection result. Here, the most reliable way to select the lambda parameter is to use an optimization algorithm. The Stochastic Gradient Descent (SGD) algorithm provides exemplary performance in many optimization problems. In this study, the lambda parameter used in INCA was selected with the SGD optimization algorithm. The pseudocode expression of this algorithm (INCA) is given in Algorithm 1.
Algorithm 1. Pseudocode of the PARC model
Inputs: features from the PARC model, labels,
Output: the selected feature (features_out)
1: features_out = INCA_algorithm (features, labels)
2: begin
3: nca = fscnca (Xtrain, ytrain, ‘sgd’, bestlambda);
4:   for i = 1 to N do
5:      search the best lambda parameter by using the NCA
6:   end for i
7:   compute feature weights with the best lambda
8:   indeces = weights (indices)
9:      for j = 1 to length (features) do
10:         if weights (j) >= threshold
11:            append j to the new indices list
12:         end if
13:      end for j
14:   features_out = fea[new_indeces]
15:   return features_out
16: end
The parameter N is the number of optimization iterations and is set to 20 (default value). The threshold value is used to eliminate features including low-weight values. For the binary and multi-class classification, threshold values were selected as 0.5 and 0.2, respectively.

8. Experimental Studies

The computer used in the experimental studies has a 4 GB graphics card, i7 intel 5500U processor (Intel, Santa Clara, CA, USA), and 16 GB RAM, and the MATLAB 2021a program was installed on Windows 11 and used to code the suggested technique. Training the suggested Parallel Attention–Residual CNN (PARC) model was carried out for multi-class and binary classification at the first step of the suggested methodology. The epoch value was chosen as 100 to achieve maximum performance. This ensured sufficient training iterations. The mini-batch size was chosen as 32. This was the maximum allowed by the hardware for the PARC model used. The initial learning rate was chosen as 0.001. For smaller values, the training time increased and for larger values, the classification performance decreased. The SGDM technique was used as the optimization solver. The validation method employed was the 10-fold cross-validation, and the loss function was the cross-entropy. It took around 2 h to train the PARC model with the available hardware.
The loss values and accuracy graphics during optimization are presented in Figure 7 and Figure 8. The training–validation accuracy scores for the binary classification reached 100% and 95.38%, and the training–validation loss values reached 0.035 and 0.1, respectively. The training–validation accuracy scores for the multi-class classification reached 99.21% and 91.89%, and the training–validation loss values reached 0.15 and 0.12, respectively.
By using learnable parameters and input data, the extraction of five hundred deep features was performed by “fc4”, which is a fully connected layer in PARC. Therefore, the SVM algorithm could be used for the classification task instead of the softmax classifier. The Iterative Neighborhood Component Analysis (INCA) algorithm selected the most distinctive features and the computational cost decreased the execution time of the SVM classifier code. The number of the nearest neighbor (hyperparameter) was chosen as 10 (default value). Figure 9 shows the computed feature weights for each feature index. The threshold weights for the features to be selected with the INCA algorithm are set to 0.5 and 0.2 for binary and multi-class classification, respectively.
For binary and multi-class classification problems, 20 and 59 distinctive features were automatically selected by the INCA algorithm. Three-dimensional representations of the selected features are given in Figure 10 and Figure 11 for two classification problems. In Figure 10 and Figure 11, it is seen that the distinguishing characteristics of the features are increased with the feature selection process. In both figures, the rows represent classes and the columns show whether the feature selection process has been performed. The x-direction in the figures indicates the number of features. The y-direction provides the features’ amplitude values while the z-direction represents the feature depth. To show the features in three dimensions, this parameter has been added. For each display, it is set to 2.
The accuracy scores according to feature selection cases and different classifiers are given in Table 1 for two classification problems. DT, SVM, KNN, NB, LD, Subspace KNN (SK), Subspace Discriminate (SD), and RUSBoosted Trees (RT) were classifier algorithms in the Classification Learner (CL) tool in the Matlab program. The default hyperparameters were selected in the CL tool for the classification process. This ablation study was performed to evaluate the feature selection operation’s effectiveness and detect which classifier algorithm gave the best accuracy.
As seen in Table 1, the SVM with the Gaussian kernel provided the best accuracy for two classification problems. In the case without feature selection, for binary and multi-class classification problems, the accuracy scores were 95.4% and 94.6%, respectively. In the case of the INCA feature selection, for binary and multi-class classification problems, the accuracy scores were 98.3% and 97.5%, respectively. Subspace Discriminant and RUSBoosted Trees provided the worst accuracies for two classification problems. In multi-class classification and binary classification without feature selection, the worst performance of the accuracy rate was 88.7% (Subspace Discriminant) and 85.6% (RUSBoosted Trees), respectively. In multi-class classification and binary classification with the INCA feature selection, the worst performances of the accuracy rate were 90.1% (Subspace Discriminant) and 88.5% (RUSBoosted Trees), respectively.
Figure 12 shows the confusion matrices of the proposed approach for the binary and multi-class classification tasks. In Figure 12a, the classes named 1, 2, 3, 4, and 5 represent the pitted surface, crazing, scratches, patches, and multi-class defect classes, respectively. In Figure 12b, the classes named 1, 2, 3, 4, and 5 represent the defect and non-defect classes, respectively. For the binary and multi-class classification problems, the accuracy scores were 98.3% and 97.5%, respectively.
Performance metrics, including the sensitivity (SN), specificity (SP), precision (PR), and F-score, were computed by using true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values. The results of the computed metrics are given in Table 2.
For the binary classification, the SN, SP, PR, and F-score values of the defect class were 0.986, 0.98, 0.983, and 0.984, respectively. The SN, SP, PR, and F-score values of the non-defect class were 0.98, 0.986, 0.984, and 0.982, respectively. For the multi-class classification, the best SN (0.992), SP (0.996), PR (0.985), and F-score (0.988) values were obtained for the scratches, patches and multi-class defects, scratches, and scratches classes, respectively. The worst SN (0.881), SP (0.96), PR (0.985), and F-scores (0.988) values were obtained for the multi-class defects, scratches, crazing, and multi-class defect classes, respectively.
Figure 13 and Figure 14 show the ROC curves and AUC values for the two classification problems. In Figure 13, positive classes 1 and 2 include the defect and non-defect classes, respectively. In Figure 14, positive classes 1, 2, 3, 4, and 5 include pitted surfaces, crazing, scratches, patches, and the multi-class defect classes, respectively.
As seen in Figure 13a,b, the AUC values were 0.99 for positive classes 1 and 2. As seen in Figure 14a–e, the AUC values were 0.99, 1.00, 0.99, 0.99, and 0.98 for positive classes 1, 2, 3, 4, and 5, respectively.

9. Discussion

In this section, ablation studies for the proposed approach were performed and the proposed approach was compared with state-of-the-art methods. Figure 15 and Figure 16 show the effect of processed images (PI) and residual attention strategies on classification accuracy. Descriptive summary information about these ablation studies is provided in Table 3 for binary and multi-class classification, respectively.
As seen in Figure 15 for the binary classification, the best accuracy was obtained by the proposed approach (PI + PARC + SVM) while the worst accuracy was obtained by the raw image (RI) + CNN strategy (Figure 15a). In Figure 15b, PI instead of RI was used for the classification and the classification accuracy was improved by 3.5%. In Figure 15c, the attention structure (PI + A-CNN) was added to the CNN model in Figure 15b. The classification accuracy was improved by 0.4%. In Figure 15d, the residual structure (PI + R-CNN) was added to the CNN model in Figure 15b. The classification accuracy was improved by 1.0%. In Figure 15e, the parallel residual and attention structure (PI + PARC) was added to the CNN model in Figure 15b. The classification accuracy was improved by 1.5%. In Figure 15f, the SVM classifier with INCA feature selection algorithm was applied in place of the softmax classifier in the proposed Parallel Attention–Residual CNN (PARC) model and the accuracy of classification was improved by 2.9% compared to the model in Figure 15e.
As seen in Figure 16 for the multi-class classification, the best accuracy was obtained by the proposed approach (PI + PARC + SVM) while the worst accuracy was obtained by the raw image (RI) + CNN strategy (Figure 16a). In Figure 16b, PI instead of RI was used for the classification, and the accuracy performance of classification was increased by 2.0%. In Figure 16c, attention structure (PI + A-CNN) was added to the CNN model in Figure 16b. The classification accuracy was improved by 0.7%. In Figure 16d, the residual structure (PI + R-CNN) was added to the CNN model in Figure 16b. The classification accuracy was improved by 2.0%. In Figure 16e, the parallel residual and attention structure (PI + PARC) was added to the CNN model in Figure 16b. The classification accuracy was improved by 2.7%. In Figure 16f, the SVM classifier with the INCA algorithm was used instead of the softmax classifier in the suggested PARC method, and the accuracy performance of classification was improved by 5.6% in comparison with the model in Figure 16e.
Table 4 presents the performance results of the proposed approach and state-of-the-art models using the dataset named “Severstal: Steel Defect Detection”. Fadli and Herlistiono [50] used the Xception network, a pre-trained CNN model for automated steel surface classification. For the binary and multi-class classification, the classification accuracies were 94% and 85%, respectively. Guo et al. [51] performed automated surface steel detection with a specific GAN model. The binary class classification performance reached 96.80% accuracy. A hybrid approach with Faster R-CNN models and ResNet50 was proposed by Wang et al. [21]. The prediction values obtained from weight activations of the ResNet model were utilized for the thresholding operation. If the scores were less than 0.3, the steel samples were considered to be defect-free. If scores were larger than 0.3, the samples were described as defective. With this method, the binary classification accuracy was 97.47%. Additionally, four copies of each steel sample image from the dataset were created, along with new class labels, using this method. The dataset was therefore multiplied by four. Steel surface flaw classification was conducted by Chigateri et al. [52] using the Exception model. They achieved an accuracy of 88% for binary classification and the same accuracy of 88% for multiple classification. A combination of the ResNet model with the squeeze-and-excitation networks suggested by Hu et al. [53] produced an accuracy value of 87.5% for the two-class classification task and 94% for the four-class classification problem.
The efficacy of the proposed method was assessed by employing an alternate dataset, specifically the NEU Surface Defect Database [54]. Consequently, the dependability of the suggested approach was enhanced. This dataset comprised six distinct types of surface defects: crazing (class 1), inclusions (class 2), patches (class 3), pitted surfaces (class 4), rolled-in scales (class 5), and scratches (class 6). Through the utilization of 10-fold cross-validation, the dataset, consisting of a total of 1800 samples distributed evenly with 300 examples in each class, underwent evaluation. Within this dataset, the suggested technique demonstrated a classification accuracy of 99.77%, as depicted in the confusion matrix presented in Figure 17.
As can be seen from the results in Table 4 and Table 5, the proposed model improved the classification performance compared to other models using the same dataset. However, given that the methods’ training parameters and training–evaluation methodology differ, this finding does not necessarily imply that the proposed strategy is better than the others.
Table 5 presents the performance results of the proposed approach and state-of-the-art models on the NEU Surface Defect Database. Yeung et al. [54] introduced a novel hybrid model, which underwent training from the ground up by incorporating the CNN architecture and five attention layers. The resultant CNN model, enriched with integrated attention, demonstrated an accuracy of 89.30%. Meanwhile, Tian et al. [55] employed the SegNet model to reconstruct and segment images of steel surfaces, and subsequent classification using a CNN model with an end-to-end learning approach yielded a success rate of 89.60%. Yi et al. [56] utilized a 14-layer CNN model featuring five convolutional layers, achieving an impressive classification accuracy of 99.05%. Li et al. [14] implemented a transfer learning system based on the ResNet model, attaining a classification accuracy of 99.00% with the ResNet CNN model. In another approach, Fu et al. [15] leveraged the SqueezeNet framework to construct a lightweight CNN model, incorporating a blur operation for raw photo processing, resulting in a remarkable success percentage of 99.61%.
The proposed approach outperformed the CNN, pre-trained CNN, R-CNN, transfer learning, and GAN models due to its novel and task-specific design. Key innovations included the use of a spectrogram algorithm in pre-processing, which transformed image data into a more detailed representation of surface defects, and the Parallel Attention–Residual CNN (PARC) model, which combined attention mechanisms for highlighting critical regions and residual connections for preserving important feature information. Unlike traditional models that relied on softmax for classification, this approach extracted deep features and evaluated them using multiple classifiers, with the SVM yielding the best results. Additionally, the INCA feature selection algorithm reduced the computational complexity while improving the classification accuracy. This tailored methodology enhanced defect detection by focusing on capturing subtle pixel variations and optimizing performance for the specific task.
Commonly used CNN models in both datasets include either pre-trained CNN models, such as ResNet, or lightly weighted CNN models. The network file sizes of the proposed PARC model and other popular pre-trained models with weights are given in Table A2. The PARC model has less weight than the other pre-trained CNN models, except the MobileNet model. Therefore, the execution time is optimal for both training and testing. In addition, accurate classification is more important than speed in the detection of steel surface defects.
Table 6 presents the classification accuracy results for both multi-class and binary classification tasks using various image enhancement techniques followed by CNN-based classification. The proposed spectrogram-based method shows superior performance in both scenarios.
The experimental results demonstrate the effectiveness of the proposed spectrogram-based method in enhancing the discriminability of surface defects for both multi-class and binary classification tasks. While the baseline approach using raw images with CNN achieved a 93.0% accuracy for multi-class and 87.2% for binary classification, applying conventional image enhancement techniques such as histogram equalization, gamma correction, CLAHE, unsharp masking, and log/power-law transforms led to modest improvements. Among these, CLAHE and histogram equalization performed relatively better due to their ability to improve the local and global contrast, respectively. However, the proposed method outperformed all the others, achieving the highest classification accuracy of 93.5% in the multi-class scenario and 89.2% in binary classification.
This superior performance can be attributed to the spectrogram’s ability to transform gradient-derived signals into the time–frequency domain, enabling the capture of both spatial patterns and frequency-based features related to surface defects. Unlike conventional techniques that primarily operate in the spatial domain and focus on intensity or contrast, the spectrogram representation provides a richer and more informative input to the CNN by highlighting subtle defect structures. These results quantitatively validate the advantage of the proposed method over traditional enhancement techniques in terms of feature extraction and defect classification performance.

10. Conclusions

This study focuses on the automatic detection of surface defects, which is an important issue in steel fabrication. It has been made to increase the automatic classification performance with a specific deep learning-based strategy. The classification performance is enhanced by processed images and feature extraction in the Parallel Attention–Residual CNN (PARC) model. The PARC model outperformed the CNN model (no residual and attention structures), the R-CNN (Residual-CNN), and the A-CNN (Attention-CNN) models. In addition, the Iterative Neighborhood Component Analysis (INCA) algorithm efficiently reduced the size of the feature set and improved the classification performance for both datasets. The classification performance values obtained for the Severstal Dataset (SD) and the Neu-Surface Dataset (NSD) are summarized in Table 7. As can be seen from Table 7, both datasets achieved 0.94 and above in all the performance metrics. In both datasets, the classification accuracy was improved according to the model that achieves the best performance out of the existing models. The classification performance is improved by 1.43%, 7.0%, and 0.16% for SD 2-class, SD 4-class, and NSD, respectively.
The good classification performance of the proposed approach makes it possible to use it for the real-time detection of steel surface defects. In the next phase, the recorded weights of the proposed approach can be tested on an artificial intelligence development kit such as an NVIDIA Jetson Orin Nano (NVIDIA, Santa Clara, CA, USA). Thus, once the test performance of the model is confirmed, it can be used in enterprises.
The most important limitation of the proposed model is that it is difficult to implement in real-time embedded systems due to the size of the model. This limitation can be solved with server-based systems. However, this may increase the financial cost.

Author Contributions

Conceptualization, F.D. and K.S.P.; methodology, F.D.; software, F.D.; validation, F.D. and K.S.P.; formal analysis, F.D. and K.S.P.; investigation, F.D. and K.S.P.; resources, F.D. and K.S.P.; data curation, F.D. and K.S.P.; writing—original draft preparation, F.D. and K.S.P.; writing—review and editing, F.D. and K.S.P.; visualization, F.D. and K.S.P.; supervision, F.D. and K.S.P.; project administration, F.D. and K.S.P.; funding acquisition, F.D. and K.S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets are publicly available on https://github.com/siddhartamukherjee/NEU-DET-Steel-Surface-Defect-Detection (accessed on 10 March 2025), https://www.kaggle.com/competitions/severstal-steel-defect-detection (accessed on 10 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Layers of the PARC model.
Table A1. Layers of the PARC model.
LayerLayer OrderLayerLayer Info
1‘conv2d_1’2D ConvolutionIn total, 16 5 × 5 × 3 convolutions with stride [1 1] and padding ‘same’
2‘batchnorm_1’Batch NormalizationBatch normalization with 16 channels
3‘relu1’ReLUReLU
4‘maxpool2d_1’2D Max PoolingIn total, 2 × 2 max pooling with stride [1 1] and padding [0 0 0 0]
5‘conv2d2_1’2D ConvolutionIn total, 8 3 × 3 × 16 convolutions with stride [1 1] and padding ‘same’
6‘batchnorm_2’Batch NormalizationBatch normalization with 8 channels
7‘relu2’ReLUReLU
8‘maxpool2d2’2D Max PoolingIn total, 2 × 2 max pooling with stride [1 1] and padding [0 0 0 0]
9‘conv2d_2’2D ConvolutionIn total, 8 2 × 2 × 8 convolutions with stride [1 1] and padding ‘same’
10‘relu_2_1_1_5’ReLUReLU
11‘conv2d_3’2D ConvolutionIn total, 8 3 × 3 × 8 convolutions with stride [1 1] and padding ‘same’
12‘maxpool2d_2’2D Max PoolingIn total, 2 × 2 max pooling with stride [1 1] and padding ‘same’
13‘conv2d2_2’2D ConvolutionIn total, 8 5 × 5 × 8 convolutions with stride [1 1] and padding ‘same’
14‘relu_2_1_1_6’ReLUReLU
15‘conv2d2_3’2D ConvolutionIn total, 8 3 × 3 × 8 convolutions with stride [1 1] and padding ‘same’
16‘sigmoid_1_1_1_3’sigmoidLayersigmoidLayer
17‘mul_1_1_1_3’ElementWiseMultiplicationElement-Wise Multiplication of 2 inputs
18‘fc0’Fully ConnectedIn total, 350 fully connected layers
19‘ReLu2’ReLUReLU
20‘input’Image InputIn total, 32 × 200 × 3 images with ‘zero center’ normalization
21‘convInp’2D ConvolutionIn total, 16 3 × 3 × 3 convolutions with stride [1 1] and padding ‘same’
22‘BNInp’Batch NormalizationBatch normalization with 16 channels
23‘reluInp’ReLUReLU
24‘S1U1conv1’2D ConvolutionIn total, 16 3 × 3 × 16 convolutions with stride [1 1] and padding ‘same’
25‘S1U1BN1’Batch NormalizationBatch normalization with 16 channels
26‘S1U1relu1’ReLUReLU
27‘S1U1conv2’2D ConvolutionIn total, 16 3 × 3 × 16 convolutions with stride [1 1] and padding ‘same’
28‘S1U1BN2’Batch NormalizationBatch normalization with 16 channels
29‘add11’AdditionElement-wise addition of 2 inputs
30‘relu11’ReLUReLU
31‘S1U2conv1’2D ConvolutionIn total, 16 3 × 3 × 16 convolutions with stride [1 1] and padding ‘same’
32‘S1U2BN1’Batch NormalizationBatch normalization with 16 channels
33‘S1U2relu1’ReLUReLU
34‘S1U2conv2’2D ConvolutionIn total, 16 3 × 3 × 16 convolutions with stride [1 1] and padding ’same’
35‘S1U2BN2’Batch NormalizationBatch normalization with 16 channels
36‘add12’AdditionElement-wise addition of 2 inputs
37‘relu12’ReLUReLU
38‘S2U1conv1’2D ConvolutionIn total, 32 3 × 3 × 16 convolutions with stride [2 2] and padding ‘same’
39‘S2U1BN1’Batch NormalizationBatch normalization with 32 channels
40‘S2U1relu1’ReLUReLU
41‘S2U1conv2’2D ConvolutionIn total, 32 3 × 3 × 32 convolutions with stride [1 1] and padding ‘same’
42‘S2U1BN2’Batch NormalizationBatch normalization with 32 channels
43‘add21’AdditionElement-wise addition of 2 inputs
44‘relu21’ReLUReLU
45‘S2U2conv1’2D ConvolutionIn total, 32 3 × 3 × 32 convolutions with stride [1 1] and padding ‘same’
46‘S2U2BN1’Batch NormalizationBatch normalization with 32 channels
47‘S2U2relu1’ReLUReLU
48‘S2U2conv2’2D ConvolutionIn total, 32 3 × 3 × 32 convolutions with stride [1 1] and padding ‘same’
49‘S2U2BN2’Batch NormalizationBatch normalization with 32 channels
50‘add22’AdditionElement-wise addition of 2 inputs
51‘relu22’ReLUReLU
52‘S3U1conv1’2D ConvolutionIn total, 64 3 × 3 × 32 convolutions with stride [2 2] and padding ‘same’
53‘S3U1BN1’Batch NormalizationBatch normalization with 64 channels
54‘S3U1relu1’ReLUReLU
55‘S3U1conv2’2D ConvolutionIn total, 64 3 × 3 × 64 convolutions with stride [1 1] and padding ‘same’
56‘S3U1BN2’Batch NormalizationBatch normalization with 64 channels
57‘add31’AdditionElement-wise addition of 2 inputs
58‘relu31’ReLUReLU
59‘S3U2conv1’2D ConvolutionIn total, 64 3 × 3 × 64 convolutions with stride [1 1] and padding ‘same’
60‘S3U2BN1’Batch NormalizationBatch normalization with 64 channels
61‘S3U2relu1’ReLUReLU
62‘S3U2conv2’2D ConvolutionIn total, 64 3 × 3 × 64 convolutions with stride [1 1] and padding ‘same’
63‘S3U2BN2’Batch NormalizationBatch normalization with 64 channels
64‘add32’AdditionElement-wise addition of 2 inputs
65‘relu32’ReLUReLU
66‘globalPool’2D Average PoolingIn total, 8 × 8 average pooling with stride [1 1] and padding [0 0 0 0]
67‘fc3’Fully ConnectedIn total, 350 fully connected layers
68‘relu_3’ReLUReLU
69‘add42’AdditionElement-wise addition of 2 inputs
70‘fc4’Fully ConnectedIn total, 150 fully connected layers
71‘relu_4’ReLUReLU
72‘fcFinal’Fully ConnectedIn total, 2 fully connected layers
73‘softmax’Softmaxsoftmax
74‘classoutput’Classification Outputcrossentropyex with classes ‘defect’ and ‘non_defect’
75‘skipConv1’2D ConvolutionIn total, 32 1 × 1 × 16 convolutions with stride [2 2] and padding [0 0 0 0]
76‘skipBN1’Batch NormalizationBatch normalization with 32 channels
77‘skipConv2’2D ConvolutionIn total, 64 1 × 1 × 32 convolutions with stride [2 2] and padding [0 0 0 0]
78‘skipBN2’Batch NormalizationBatch normalization with 64 channels
Table A2. Sizes of the PARC model and popular pre-trained CNN models.
Table A2. Sizes of the PARC model and popular pre-trained CNN models.
CNN ModelModel File Size (MB)
ResNet93.2
VGGNet502.9
AlexNet222.03
MobileNet12.95
PARC Model64.7
Table A3. Pseudocode for the implementation steps of the proposed method.
Table A3. Pseudocode for the implementation steps of the proposed method.
1. Begin with raw images containing surface defects.
2. Stage 1: Pre-processing
  a. Apply spectrogram algorithm to the time series signals of the raw images.
  b. Convert 1D data using all pixel values from the images.
  c. Obtain spectrogram images representing pixel changes in defect regions.
3. Stage 2: Model Training with PARC (Parallel Attention–Residual CNN)
  a. Design and initialize the PARC model.
    i. Combine attention and residual modules with a customized CNN.
    ii. Train attention module to highlight important image regions.
    iii. Use residual module to pass feature maps from earlier layers to later convolutional layers.
  b. Train the PARC model with the processed spectrogram images.
4. Stage 3: Feature Extraction
  a. Extract deep features from the fully connected (FC) layer of the PARC model.
  b. Use these features (instead of the softmax classifier) to test with other classifier algorithms.
5. Stage 4: Feature Selection
  a. Apply the INCA feature selection algorithm.
  b. Reduce computational cost and improve classification performance by selecting the most relevant features.
6. Stage 5: Classification
  a. Train seven popular classifiers on the selected features:
    i. Decision Tree (DT).
    ii. Support Vector Machine (SVM).
    iii. K-Nearest Neighbor (KNN).
    iv. Naïve Bayes (NB).
    v. Linear Discriminant (LD).
    vi. Subspace KNN.
    vii. Subspace Discriminate.
    viii. RUSBoosted Trees.
  b. Evaluate the classification performance of each algorithm.
  c. Select the best-performing algorithm (SVM) for the final classification.
7. End

References

  1. Zhang, D.; Song, K.; Xu, J.; He, Y.; Niu, M.; Yan, Y. MCnet: Multiple Context Information Segmentation Network of No-Service Rail Surface Defects. IEEE Trans. Instrum. Meas. 2021, 70, 5004309. [Google Scholar] [CrossRef]
  2. Hanbay, K.; Talu, M.F.; Özgüven, Ö.F. Fabric Defect Detection Systems and Methods—A Systematic Literature Review. Optik 2016, 127, 11960–11973. [Google Scholar] [CrossRef]
  3. Zhao, W.; Chen, F.; Huang, H.; Li, D.; Cheng, W. A New Steel Defect Detection Algorithm Based on Deep Learning. Comput. Intell. Neurosci. 2021, 2021, 5592878. [Google Scholar] [CrossRef] [PubMed]
  4. Dong, H.; Song, K.; He, Y.; Xu, J.; Yan, Y.; Meng, Q. PGA-Net: Pyramid Feature Fusion and Global Context Attention Network for Automated Surface Defect Detection. IEEE Trans. Ind. Inform. 2020, 16, 7448–7458. [Google Scholar] [CrossRef]
  5. Cao, J.; Yang, G.; Yang, X. A Pixel-Level Segmentation Convolutional Neural Network Based on Deep Feature Fusion for Surface Defect Detection. IEEE Trans. Instrum. Meas. 2021, 70, 5003712. [Google Scholar] [CrossRef]
  6. Dikbas, H.; Taskaya, S. Alloying the Surface of AISI 2205 Duplex Stainless Steel Material by PTA Welding Method and Making Its Thermomechanical Investigation in ANSYS Software. J. Therm. Anal. Calorim. 2020, 139, 3847–3856. [Google Scholar] [CrossRef]
  7. Konovalenko, I.; Maruschak, P.; Brezinová, J.; Viňáš, J.; Brezina, J. Steel Surface Defect Classification Using Deep Residual Neural Network. Metals 2020, 10, 846. [Google Scholar] [CrossRef]
  8. den Bakker, I. Python Deep Learning Cookbook: Over 75 Practical Recipes on Neural Network Modeling, Reinforcement Learning, and Transfer Learning Using Python; Packt Publishing Ltd.: Birmingham, UK, 2007; Volume 136, ISBN 9781787125193. [Google Scholar]
  9. Martins, L.A.O.; Pádua, F.L.C.; Almeida, P.E.M. Automatic Detection of Surface Defects on Rolled Steel Using Computer Vision and Artificial Neural Networks. In Proceedings of the IECON Proceedings (Industrial Electronics Conference), Glendale, Arizona, USA, 7–10 November 2010; IEEE: New York, NY, USA, 2010; pp. 1081–1086. [Google Scholar]
  10. Li, Z.; Wei, X.; Hassaballah, M.; Li, Y.; Jiang, X. A Deep Learning Model for Steel Surface Defect Detection. Complex Intell. Syst. 2024, 10, 885–897. [Google Scholar] [CrossRef]
  11. Pang, W.; Tan, Z. A Steel Surface Defect Detection Model Based on Graph Neural Networks. Meas. Sci. Technol. 2024, 35, 46201. [Google Scholar] [CrossRef]
  12. Zhang, H.; Li, S.; Miao, Q.; Fang, R.; Xue, S.; Hu, Q.; Hu, J.; Chan, S. Surface Defect Detection of Hot Rolled Steel Based on Multi-Scale Feature Fusion and Attention Mechanism Residual Block. Sci. Rep. 2024, 14, 7671. [Google Scholar] [CrossRef]
  13. Zhao, Y.J.; Yan, Y.H.; Song, K.C. Vision-Based Automatic Detection of Steel Surface Defects in the Cold Rolling Process: Considering the Influence of Industrial Liquids and Surface Textures. Int. J. Adv. Manuf. Technol. 2017, 90, 1665–1678. [Google Scholar] [CrossRef]
  14. Li, J.; Su, Z.; Geng, J.; Yin, Y. Real-Time Detection of Steel Strip Surface Defects Based on Improved YOLO Detection Network. IFAC-PapersOnLine 2018, 51, 76–81. [Google Scholar] [CrossRef]
  15. Fu, G.; Sun, P.; Zhu, W.; Yang, J.; Cao, Y.; Yang, M.Y.; Cao, Y. A Deep-Learning-Based Approach for Fast and Robust Steel Surface Defects Classification. Opt. Lasers Eng. 2019, 121, 397–405. [Google Scholar] [CrossRef]
  16. Liu, Y.; Xu, K.; Xu, J. Periodic Surface Defect Detection in Steel Plates Based on Deep Learning. Appl. Sci. 2019, 9, 3127. [Google Scholar] [CrossRef]
  17. Liu, Y.; Yuan, Y.; Balta, C.; Liu, J. A Light-Weight Deep-Learning Model with Multi-Scale Features for Steel Surface Defect Classification. Materials 2020, 13, 4629. [Google Scholar] [CrossRef]
  18. Guan, S.; Lei, M.; Lu, H. A Steel Surface Defect Recognition Algorithm Based on Improved Deep Learning Network Model Using Feature Visualization and Quality Evaluation. IEEE Access 2020, 8, 49885–49895. [Google Scholar] [CrossRef]
  19. Amin, D.; Akhter, S. Deep Learning-Based Defect Detection System in Steel Sheet Surfaces. In Proceedings of the 2020 IEEE Region 10 Symposium, TENSYMP 2020, Dhaka, Bangladesh, 5–7 June 2020; IEEE: New York, NY, USA, 2020; pp. 444–448. [Google Scholar]
  20. Severstal: Steel Defect Detection. Available online: https://www.kaggle.com/c/severstal-steel-defect-detection/overview (accessed on 10 March 2025).
  21. Wang, S.; Xia, X.; Ye, L.; Yang, B. Automatic Detection and Classification of Steel Surface Defect Using Deep Convolutional Neural Networks. Metals 2021, 11, 388. [Google Scholar] [CrossRef]
  22. Demir, F.; Bajaj, V.; Ince, M.C.; Taran, S.; Şengür, A. Surface EMG Signals and Deep Transfer Learning-Based Physical Action Classification. Neural Comput. Appl. 2019, 31, 8455–8462. [Google Scholar] [CrossRef]
  23. Özseven, T. Investigation of the Effect of Spectrogram Images and Different Texture Analysis Methods on Speech Emotion Recognition. Appl. Acoust. 2018, 142, 70–77. [Google Scholar] [CrossRef]
  24. Kollias, D.; Zafeiriou, S. Exploiting Multi-CNN Features in CNN-RNN Based Dimensional Emotion Recognition on the OMG in-the-Wild Dataset. IEEE Trans. Affect. Comput. 2021, 12, 595–606. [Google Scholar] [CrossRef]
  25. Demir, K.; Berna, A.R.I.; Demir, F. Detection of Brain Tumor with a Pre-Trained Deep Learning Model Based on Feature Selection Using MR Images. Firat Univ. J. Exp. Comput. Eng. 2023, 2, 23–31. [Google Scholar] [CrossRef]
  26. Canan, K.O.Ç.; Özyurt, F. An Examination of Synthetic Images Produced with DCGAN According to the Size of Data and Epoch. Firat Univ. J. Exp. Comput. Eng. 2023, 2, 32–37. [Google Scholar]
  27. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  28. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning ICML 2015, Lille, France, 7–9 July 2015; Volume 1, pp. 448–456. [Google Scholar]
  30. Santurkar, S.; Tsipras, D.; Ilyas, A.; Mądry, A. How Does Batch Normalization Help Optimization? In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 2–8 December 2018; pp. 2488–2498. [Google Scholar]
  31. Agarap, A.F. Deep Learning Using Rectified Linear Units (Relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  32. Weng, L.; Zhang, H.; Chen, H.; Song, Z.; Hsieh, C.-J.; Daniel, L.; Boning, D.; Dhillon, I. Towards Fast Computation of Certified Robustness for Relu Networks. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 5276–5285. [Google Scholar]
  33. Yilmaz, T.; Aydogmus, O. Deep Deterministic Policy Gradient Reinforcement Learning for Collision-Free Navigation of Mobile Robots in Unknown Environments. Firat Univ. J. Exp. Comput. Eng. 2023, 2, 87–96. [Google Scholar] [CrossRef]
  34. Liang, X.; Wang, X.; Lei, Z.; Liao, S.; Li, S.Z. Soft-Margin Softmax for Deep Classification. In Proceedings of the International Conference on Neural Information Processing, Guangzhou, China, 14–18 November 2017; pp. 413–421. [Google Scholar]
  35. Zang, F.; Zhang, J. Softmax Discriminant Classifier. In Proceedings of the 2011 Third International Conference on Multimedia Information Networking and Security, Shanghai, China, 4–6 November 2011; pp. 16–19. [Google Scholar]
  36. Akpinar, E.K.; Mehmet, D.A.Ş. Modeling of a Solar Air Collector Heat Transfer Coefficient with Regression Algorithms. Firat Univ. J. Exp. Comput. Eng. 2022, 1, 14–23. [Google Scholar] [CrossRef]
  37. Atila, O.; Şengür, A. Attention Guided 3D CNN-LSTM Model for Accurate Speech Based Emotion Recognition. Appl. Acoust. 2021, 182, 108260. [Google Scholar] [CrossRef]
  38. Niu, Z.; Zhong, G.; Yu, H. A Review on the Attention Mechanism of Deep Learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  39. Abdelaziz Ismael, S.A.; Mohammed, A.; Hefny, H. An Enhanced Deep Learning Approach for Brain Cancer MRI Images Classification Using Residual Networks. Artif. Intell. Med. 2020, 102, 101779. [Google Scholar] [CrossRef]
  40. Zhao, Y.; Liu, Y.; Huang, W. Prediction Model of HBV Reactivation in Primary Liver Cancer—Based on NCA Feature Selection and SVM Classifier with Bayesian and Grid Optimization. In Proceedings of the 2018 IEEE 3rd International Conference on Cloud Computing and Big Data Analysis (ICCCBDA), Chengdu, China, 20–22 April 2018; IEEE: New York, NY, USA, 2018; pp. 547–551. [Google Scholar]
  41. Muqeet, H.A.; Liaqat, R.; Hussain, A.; Sajjad, I.A.; Ahmad, H.; Mehmood, A. Regularized NCA Based Prominent Feature Selection for Load Identification Using Boosted Tree Classifier. 2024. Available online: https://assets-eu.researchsquare.com/files/rs-4151725/v1_covered_d7165e06-8243-4a13-81e3-1fe6a5ad9d35.pdf (accessed on 1 March 2025).
  42. Khatri, S.; Bansal, P. Hyperparameter Tuning and Validation of Neural Network Model for Software Effort Estimation with NCA Based Feature Selection. In Proceedings of the 2024 IEEE International Conference on Information Technology, Electronics and Intelligent Communication Systems (ICITEICS), Bangalore, India, 28–29 June 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  43. Demir, F.; Siddique, K.; Alswaitti, M.; Demir, K.; Sengur, A. A Simple and Effective Approach Based on a Multi-Level Feature Selection for Automated Parkinson’s Disease Detection. J. Pers. Med. 2022, 12, 55. [Google Scholar] [CrossRef]
  44. Yildiz, A.M.; Gun, M.V.; Yildirim, K.; Keles, T.; Dogan, S.; Tuncer, T.; Acharya, U.R. SKLBP14: A New Textural Environmental Sound Classification Model Based on a Squarekernelled Local Binary Pattern. Firat Univ. J. Exp. Comput. Eng. 2023, 2, 46–54. [Google Scholar] [CrossRef]
  45. Baygin, M.; Yaman, O.; Tuncer, T.; Dogan, S.; Barua, P.D.; Acharya, U.R. Automated Accurate Schizophrenia Detection System Using Collatz Pattern Technique with EEG Signals. Biomed. Signal Process. Control 2021, 70, 102936. [Google Scholar] [CrossRef]
  46. Tuncer, T.; Dogan, S.; Subasi, A. EEG-Based Driving Fatigue Detection Using Multilevel Feature Extraction and Iterative Hybrid Feature Selection. Biomed. Signal Process. Control 2021, 68, 102591. [Google Scholar] [CrossRef]
  47. Turkoglu, M. COVIDetectioNet: COVID-19 Diagnosis System Based on X-Ray Images Using Features Selected from Pre-Learned Deep Features Ensemble. Appl. Intell. 2021, 51, 1213–1226. [Google Scholar] [CrossRef]
  48. Tuncer, T.; Ertam, F. Neighborhood Component Analysis and ReliefF Based Survival Recognition Methods for Hepatocellular Carcinoma. Phys. A Stat. Mech. Its Appl. 2020, 540, 123143. [Google Scholar] [CrossRef]
  49. Demir, F.; Tașcı, B. An Effective and Robust Approach Based on R-CNN + LSTM Model and NCAR Feature Selection for Ophthalmological Disease Detection from Fundus Images. J. Pers. Med. 2021, 11, 1276. [Google Scholar] [CrossRef]
  50. Fadli, V.F.; Herlistiono, I.O. Steel Surface Defect Detection Using Deep Learning. Int. J. Innov. Sci. Res. Technol. 2020, 5, 244–250. [Google Scholar] [CrossRef]
  51. Guo, X.; Liu, X.; Królczyk, G.; Sulowicz, M.; Glowacz, A.; Gardoni, P.; Li, Z. Damage Detection for Conveyor Belt Surface Based on Conditional Cycle Generative Adversarial Network. Sensors 2022, 22, 3485. [Google Scholar] [CrossRef]
  52. Chigateri, K.B.; Hebbale, A.M. A Steel Surface Defect Detection Model Using Machine Learning. Mater. Today Proc. 2023, 100, 51–58. [Google Scholar] [CrossRef]
  53. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  54. Yeung, C.C.; Lam, K.M. Efficient Fused-Attention Model for Steel Surface Defect Detection. IEEE Trans. Instrum. Meas. 2022, 71, 2510011. [Google Scholar] [CrossRef]
  55. Tian, S.; Huang, P.; Ma, H.; Wang, J.; Zhou, X.; Zhang, S.; Zhou, J.; Huang, R.; Li, Y. CASDD: Automatic Surface Defect Detection Using a Complementary Adversarial Network. IEEE Sens. J. 2022, 22, 19583–19595. [Google Scholar] [CrossRef]
  56. Yi, L.; Li, G.; Jiang, M. An End-to-End Steel Strip Surface Defects Recognition System Based on Convolutional Neural Networks. Steel Res. Int. 2017, 88, 176–187. [Google Scholar] [CrossRef]
Figure 1. Dataset samples of classified shapes with defects: (a) pitted surface, (b) crazing, (c) scratches, (d) patches, (e) non-defect [21].
Figure 1. Dataset samples of classified shapes with defects: (a) pitted surface, (b) crazing, (c) scratches, (d) patches, (e) non-defect [21].
Applsci 15 04255 g001
Figure 2. The framework of the proposed methodology.
Figure 2. The framework of the proposed methodology.
Applsci 15 04255 g002
Figure 3. The layer structure of the proposed PARC model.
Figure 3. The layer structure of the proposed PARC model.
Applsci 15 04255 g003
Figure 4. One-dimensional stack data with the gradient operation for each class.
Figure 4. One-dimensional stack data with the gradient operation for each class.
Applsci 15 04255 g004
Figure 5. Conversion samples from raw images to spectrogram images for each class.
Figure 5. Conversion samples from raw images to spectrogram images for each class.
Applsci 15 04255 g005
Figure 6. The illustration for the proposed attention layer.
Figure 6. The illustration for the proposed attention layer.
Applsci 15 04255 g006
Figure 7. Accuracy and loss graphs of the PARC model during the training process for binary classification: (a) accuracy graph, (b) loss graph.
Figure 7. Accuracy and loss graphs of the PARC model during the training process for binary classification: (a) accuracy graph, (b) loss graph.
Applsci 15 04255 g007
Figure 8. Accuracy and loss graphs of the PARC model during the training process for multi–class classification: (a) accuracy graph, (b) loss graph.
Figure 8. Accuracy and loss graphs of the PARC model during the training process for multi–class classification: (a) accuracy graph, (b) loss graph.
Applsci 15 04255 g008
Figure 9. Features weighted with the INCA algorithm: (a) binary classification and (b) multi–class classification.
Figure 9. Features weighted with the INCA algorithm: (a) binary classification and (b) multi–class classification.
Applsci 15 04255 g009
Figure 10. Three–dimensional feature representation for no feature selection and INCA feature selection cases (binary classification).
Figure 10. Three–dimensional feature representation for no feature selection and INCA feature selection cases (binary classification).
Applsci 15 04255 g010
Figure 11. Three–dimensional feature representation for no feature selection and INCA feature selection cases (multi–class classification).
Figure 11. Three–dimensional feature representation for no feature selection and INCA feature selection cases (multi–class classification).
Applsci 15 04255 g011
Figure 12. Confusion matrices of the proposed approach: (a) multi-class classification (classes: (1) pitted surfaces, (2) crazing, (3) scratches, (4) patches, (5) multi-class defect), (b) binary classification (1: defect 2: non-defect).
Figure 12. Confusion matrices of the proposed approach: (a) multi-class classification (classes: (1) pitted surfaces, (2) crazing, (3) scratches, (4) patches, (5) multi-class defect), (b) binary classification (1: defect 2: non-defect).
Applsci 15 04255 g012
Figure 13. ROC curves and AUC values of the proposed approach: binary classification ((a) class 1: defect, (b) class 2: non-defect).
Figure 13. ROC curves and AUC values of the proposed approach: binary classification ((a) class 1: defect, (b) class 2: non-defect).
Applsci 15 04255 g013
Figure 14. ROC curves and AUC values of the proposed approach: multi-class classification (classes: (a) pitted surfaces, (b) crazing, (c) scratches, (d) patches, (e) multi-class defects).
Figure 14. ROC curves and AUC values of the proposed approach: multi-class classification (classes: (a) pitted surfaces, (b) crazing, (c) scratches, (d) patches, (e) multi-class defects).
Applsci 15 04255 g014
Figure 15. Confusion matrices for ablation studies of the proposed approach (binary classification (classes 1: defect 2: non-defect)).
Figure 15. Confusion matrices for ablation studies of the proposed approach (binary classification (classes 1: defect 2: non-defect)).
Applsci 15 04255 g015
Figure 16. Confusion matrices for ablation studies of the proposed approach (multi-class classification (classes: (1) pitted surfaces, (2) crazing, (3) scratches, (4) patches, (5) multi-class defects)).
Figure 16. Confusion matrices for ablation studies of the proposed approach (multi-class classification (classes: (1) pitted surfaces, (2) crazing, (3) scratches, (4) patches, (5) multi-class defects)).
Applsci 15 04255 g016
Figure 17. Confusion matrix results of the proposed approach on the NEU Surface Defect Database.
Figure 17. Confusion matrix results of the proposed approach on the NEU Surface Defect Database.
Applsci 15 04255 g017
Table 1. Accuracy variations according to feature selection cases and different classifiers.
Table 1. Accuracy variations according to feature selection cases and different classifiers.
ClassifierNo Feature SelectionINCA Feature Selection
BinaryMulti–ClassBinaryMulti–Class
DT90.288.791.390
SVM95.494.698.397.5
KNN92.593.195.194.9
NB93.287.995.190
LD92.893.594.195,2
SK90.192.892.395.6
SD88.793.690.195.4
RT89.385.691.888.5
Table 2. Other performance metric results of the proposed approach for binary and multi-class classification.
Table 2. Other performance metric results of the proposed approach for binary and multi-class classification.
Classification ModeClassesSNSPPRF-Score
BinaryDefect0.9860.980.9830.984
Non-defect0.980.9860.9840.982
Multi-ClassPitted surfaces0.9610.9950.9630.962
Crazing0.9380.9970.910.924
Scratches0.9920.960.9850.988
Patches0.9240.9960.9520.938
Multi-class defects0.8810.9960.9330.906
Table 3. Summarizing information about the ablation study for the proposed approach ((a) binary classification, (b) multi-class classification).
Table 3. Summarizing information about the ablation study for the proposed approach ((a) binary classification, (b) multi-class classification).
Model NameModel InfoAcc (a)Acc (b)
RI + CNNRaw images + PARC model without attention and residual structures93.00%87.20%
PI + CNNProcessed images + PARC model without attention and residual structures93.50%89.20%
PI + A-CNNProcessed images + PARC model without residual structures93.90%89.90%
PI + R-CNNProcessed images + PARC model without attention structures94.50%91.20%
PI + PARCProcessed images + PARC model (no feature selection with INCA and no SVM classifier)95.40%91.90%
PI + PARC + SVMProcessed images + PARC model +SVM (no feature selection with INCA)96.80%93.90%
PI + PARC + INCA + SVMProposed approach98.30%97.50%
Table 4. The classification accuracies of studies and the proposed model using the same dataset.
Table 4. The classification accuracies of studies and the proposed model using the same dataset.
AuthorsMethodologyAccuracy (%)
BinaryMulti-Class
Proposed modelPI + PARC + INCA + SVM98.994.5
Wang et al. [21]ResNet50 + Faster R-CNN97.47-
Guo et al. [51]GAN model96.8-
Chigateri et al. [52]Exception model87.685.0
Hu et al. [53]SEResNET5094.087.5
Fadli et al. [50]Transfer learning (Xception)94.085.0
Table 5. The classification accuracies of studies and the proposed model using the NEU Surface Defect Database.
Table 5. The classification accuracies of studies and the proposed model using the NEU Surface Defect Database.
AuthorMethodAccuracy (%)
Yeung et al. [54]Fused Attention CNN model89.3
Tian et al. [55]SegNet + CNN89.6
Yi et al. [56]Deep CNN99.05
Li et al. [14]Transfer learning (ResNet)99.0
Fu et al. [15]Lightweight CNN99.61
Proposed approachPARC, NRMI, SVM99.77
Table 6. Comparison of the proposed pre-processing method with basic image enhancement algorithms.
Table 6. Comparison of the proposed pre-processing method with basic image enhancement algorithms.
MethodDescriptionMulti-Class Accuracy (%)Binary Accuracy (%)
Raw Images + CNNBaseline method without any pre-processing.9387.2
Histogram Equalization + CNNEnhances global contrast by redistributing image intensities.93.488.7
Gamma CorrectionAdjusts brightness using nonlinear intensity mapping.93.287.9
CLAHE + CNNApplies adaptive histogram equalization locally to improve contrast in homogeneous areas.93.388.9
Unsharp MaskingSharpens the image by emphasizing edges and details.9387.5
Log/Power-Law TransformEnhances low-intensity pixels, useful for improving contrast in dark regions.93.288.6
Table 7. Summary of the experimental results of the proposed method.
Table 7. Summary of the experimental results of the proposed method.
CriteriaSD 2-ClassSD 4-ClassNSD
Accuracy (%)98.994.599.7
Sensitivity0.980.940.99
Specificity0.980.980.99
Precision0.980.940.98
F-score0.980.940.98
Performance improvement (%)1.437.00.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Demir, F.; Parlak, K.S. Increasing the Classification Achievement of Steel Surface Defects by Applying a Specific Deep Strategy and a New Image Processing Approach. Appl. Sci. 2025, 15, 4255. https://doi.org/10.3390/app15084255

AMA Style

Demir F, Parlak KS. Increasing the Classification Achievement of Steel Surface Defects by Applying a Specific Deep Strategy and a New Image Processing Approach. Applied Sciences. 2025; 15(8):4255. https://doi.org/10.3390/app15084255

Chicago/Turabian Style

Demir, Fatih, and Koray Sener Parlak. 2025. "Increasing the Classification Achievement of Steel Surface Defects by Applying a Specific Deep Strategy and a New Image Processing Approach" Applied Sciences 15, no. 8: 4255. https://doi.org/10.3390/app15084255

APA Style

Demir, F., & Parlak, K. S. (2025). Increasing the Classification Achievement of Steel Surface Defects by Applying a Specific Deep Strategy and a New Image Processing Approach. Applied Sciences, 15(8), 4255. https://doi.org/10.3390/app15084255

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop