Next Article in Journal
A Physical-Enhanced Spatio-Temporal Graph Convolutional Network for River Flow Prediction
Next Article in Special Issue
Multi-Scale Context Fusion Method with Spatial Attention for Accurate Crop Disease Detection
Previous Article in Journal
Design and Aerodynamic Analysis of Rigid Wing Sail of Unmanned Sailboat at Sea Based on CFD
Previous Article in Special Issue
An Improved Lightweight Model for Defect Detection on Paths in Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Fusion Using Deep Learning Algorithms in Image Classification for Security Purposes by Random Weight Network

1
Department of Computer Engineering, Faculty of Computer and Information Sciences, Konya Technical University, Konya 42250, Türkiye
2
Department of Software Engineering, Faculty of Engineering and Natural Sciences, Gümüşhane University, Gümüşhane 29100, Türkiye
3
Department of Software Engineering, Faculty of Computer and Information Sciences, Konya Technical University, Konya 42250, Türkiye
4
College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(16), 9053; https://doi.org/10.3390/app15169053
Submission received: 1 July 2025 / Revised: 13 August 2025 / Accepted: 14 August 2025 / Published: 17 August 2025
(This article belongs to the Special Issue Deep Learning for Image Processing and Computer Vision)

Abstract

Automated threat detection in X-ray security imagery is a critical yet challenging task, where conventional deep learning models often struggle with low accuracy and overfitting. This study addresses these limitations by introducing a novel framework based on feature fusion. The proposed method extracts features from multiple and diverse deep learning architectures and classifies them using a Random Weight Network (RWN), whose hyperparameters are optimized for maximum performance. The results show substantial improvements at each stage: while the best standalone deep learning model achieved a test accuracy of 83.55%, applying the RWN to a single feature set increased accuracy to 94.82%. Notably, the proposed feature fusion framework achieved a state-of-the-art test accuracy of 97.44%. These findings demonstrate that a modular approach combining multi-model feature fusion with an efficient classifier is a highly effective strategy for improving the accuracy and generalization capability of automated threat detection systems.

1. Introduction

The integrity of security screening in public spaces, such as airports, is a critical component of both national and international safety. Although numerous security measures are implemented, systems reliant on human operators remain vulnerable to error, potentially leading to severe security breaches with significant material and societal consequences. X-ray imaging systems play a central role in these security protocols, particularly for baggage inspection. However, the manual identification of concealed threats within complex environments, such as improvised explosive circuits hidden inside electronic devices like laptops, presents a formidable challenge. This specific task requires a high level of specialized expertise and is inherently prone to human oversight, thereby creating a significant vulnerability in security checkpoints.
In response to these challenges, automated detection systems driven by deep learning have been explored by researchers. However, early investigations that applied conventional deep learning models directly to this problem revealed significant limitations. These limitations include relatively low classification accuracy and a high tendency towards overfitting, largely attributed to the complexity and inherent variations within the X-ray dataset [1]. The overlapping nature of components in X-ray images increases intra-class variation, while the visual similarity between benign laptop circuits and threat items elevates inter-class confusion, making robust classification difficult. To overcome these deficiencies, this study proposes a novel framework centered on feature fusion combined with a Random Weight Network (RWN) for classification. The core hypothesis is that features extracted from multiple and diverse deep learning architectures can provide a richer, more discriminative representation of the input data. By fusing these features and employing an RWN, which is noted for its rapid training and resistance to overfitting, it is anticipated that a more accurate and generalizable classification model can be achieved. This approach addresses the key research questions regarding the performance enhancement that can be achieved through feature fusion and the optimal configuration of the RWN classifier, including the impact of hidden neuron count and activation function.
The main contributions of this work are systematically outlined as follows:
(a)
A Novel Feature Fusion Framework: This study proposes and validates a new framework that integrates features extracted from multiple deep learning models (e.g., ShuffleNet, InceptionV3) and employs a Random Weight Network (RWN) for classification. This multi-source feature fusion strategy marks a significant departure from conventional single-model approaches.
(b)
Significant Performance Improvement: A substantial improvement in classification performance is demonstrated. The proposed feature fusion methodology achieves a test accuracy of 97.44%. This result is markedly superior to both the 83.55% accuracy of the best-performing individual deep learning model, ShuffleNet, and the 94.82% accuracy from classification using features from a single model with an optimized RWN.
(c)
Comprehensive Empirical Analysis: A comprehensive empirical analysis of the RWN-based classifier is conducted. The investigation evaluates the influence of critical hyperparameters, including the number of hidden neurons and the choice of activation functions, providing a clear optimization guide for similar security applications.
(d)
Robustness and Generalization: The robustness and generalization capability of the proposed method are established through a comparative analysis against 11 state-of-the-art machine learning classifiers. The framework is shown to offer superior generalization and effective mitigation of overfitting.
(e)
Publicly Available Dataset: A challenging new dataset of X-ray images, featuring laptops with and without concealed circuits, has been created and made publicly available [1], thereby providing a valuable benchmark for future research in this domain.
Within this framework, the following research questions are posed to articulate the study’s core contributions and key capabilities:
(a)
How does it affect the classification performance using RWN on datasets whose features are extracted by deep learning models?
(b)
Can the combination of features extracted from different deep learning models significantly improve training and test accuracy in classification?
(c)
What are the performance implications of existing deep learning algorithms when applied to X-ray security datasets, and how can these be addressed through feature fusion techniques?
(d)
How does the use of an RWN influence classification performance when compared to standard deep learning models on X-ray datasets?
(e)
Do the combinations of merged features (e.g., N|M and M|N) have a significant effect on classification outcomes in RWN?
(f)
What is the impact of the number of hidden layer neurons on the performance of an RWN, and how can the risk of overfitting be minimized through optimal parameter selection?
(g)
How does the selection of activation functions (sigmoid, tangent sigmoid, or hardlim) affect the classification performance of an RWN, particularly in the context of combined datasets?
The organization of the study is as follows: In Section 1, an introduction to the study is provided, a literature review is presented, and the motivation and contribution of the study are outlined. Section 2 covers feature extraction from deep learning models, feature fusion, and dataset explanation. In Section 3, experiments are conducted, and the results obtained are analyzed. Section 4 discusses the findings, and the study is finally concluded in Section 5.
X-ray imaging technologies have been used in various aspects of daily life, as well as in fields such as crystallography, astronomy, and medicine, since the discovery of X-rays by Wilhelm Conrad Rontgen. These technologies encompass a wide range of purposes and methods, including traditional transmission methods, dual-energy techniques, and scattered X-ray methods [2]. In these technologies, rays emitted from an X-ray source are attenuated as they pass through objects. This decrease in intensity is utilized to calculate the density (d) and effective atomic number (Zeff) of the materials [3]. Consequently, materials with higher density, which cause greater attenuation, appear brighter in X-ray images, while lower-density materials appear darker. X-ray technologies are widely used for various purposes, as evidenced by the information provided by X-ray devices. Applications range from inspecting welds in industrial settings and identifying bone fractures in medicine to detecting prohibited materials in security-sensitive locations like airports, courthouses, and shopping malls.
X-ray images are utilized for the detection of prohibited materials, aiming to minimize security risks at airports through the application of machine learning and image processing techniques. This involves identifying items passengers are forbidden to carry, whether on their person or in their luggage, by analyzing 3D or 2D X-ray images [4]. These applications are typically employed to assist personnel conducting baggage control or to automate the process. This section reviews the literature on X-ray image analysis and feature fusion using deep learning algorithms.
Previously, tasks such as classification in X-ray imaging were performed using manually extracted features, such as SIFT and PHOW, often within a Bag of Words (BoW) framework [5].
In later periods, the success of convolutional neural network (CNN) techniques led to their increased use in this field as well. Akçay et al. [6] implemented transfer learning in CNN using the fine-tuning paradigm. Jaccard et al. [7] detected the presence of threat materials in cargo containers using CNN on image patches. Mery et al. [8] compared methods such as Bag of Words, Sparse Representations, deep learning, and classical pattern recognition schemes. Jaccard et al. [9] also detected cars within cargo using CNN with data augmentation. Rogers et al. [10] used the original dual-energy images as separate channels in their CNN. They performed data augmentation with Threat Image Projection. Caldwell et al. [11], investigated transfer learning in different scenarios using deep networks, such as VGG. Morris et al. [12] focused on threat detection of traditional explosives using CNNs like VGG and Inception. In addition to these, newly emerged CNN models such as region-based CNNs [13] and single-shot models like YOLO (You Look Only Once) [14] have also been applied in X-ray imaging. Petrozziello and Jordanov [15] performed the detection of steel barrel holes using CNN and Stacked Autoencoder. Cheng et al. [16] used a YOLO-based method they called X-YOLO, which has feature fusion and attention mechanisms. Wu and Xu [17] used a hybrid Self-Supervised Learning in the pre-training phase to perform detection with a structure containing a transformer in the final stage of the Head-Tail Feature Pyramid head. Wang et al. [18] used a Yolov8-based method with a dual branch structure that includes Sobel convolution and convolution branches. Additionally, in the fusion part, they used a lightweight star operation module. In addition to classification, deep learning methods have also been employed for data augmentation. Yang et al. [19] performed data augmentation using generative adversarial networks (GANs), which utilize Fréchet Inception Distance scores and compared them (DCGAN, WGAN-GP). Kaminetzky and Mery [20] performed data augmentation using simulated 3D X-ray image models. Apart from these tasks, CNNs have also been used as feature extractors. Caldwell and Griffin [21] performed data augmentation with photographic images, using both photographs and X-ray images of the same object. Benedykciuk et al. [22] addressed the material recognition problem using a multiscale network structure consisting of five subnetworks and using image patches. Babalik and Babadag [23] used CNNs as feature extractors; in the next stages, they selected the features with a binary Sparrow Search Algorithm and classified them using Support Vector Machines (SVM) and k-nearest neighbors (KNN) as classifiers. Methods such as ensemble learning and feature fusion have also been used to improve the performance of deep learning models. Ayantayo et al. [24] proposed three different deep learning models, which used early-fusion, late-fusion, and late-ensemble learning strategies to resist overfitting. Zhang et al. [25] used multi-domain features, employing transfer learning and feature fusion. They used SVM in feature extraction and model selection stages, then fused these features and performed baby cry detection. Wu et al. [26] performed deep learning-based fault diagnosis for rolling bearings. In their method, they used a new multiscale feature fusion deep residual network containing multiple multiscale feature fusion blocks and a multiscale pooling layer. Liu et al. [27] implemented a multi-modal fusion approach that combines two different deep learning models trained on simple clinical information and group images, using logistic regression for breast nodule diagnosis. Patil and Kirange [28] proposed a method for detecting brain tumors by fusing deep network features, such as VGG Inception and a shallow CNN. Gill et al. [29] used deep learning methods including CNN, Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) for image-based classification of fruits using early fusion and late fusion strategies. Deng et al. [30] addressed side-channel attacks in information security using feature extraction with a multi-scale feature fusion mechanism. Al-Timemy et al. [31] utilized Xception and InceptionResNetV2 deep learning architectures to extract features from three different corneal maps. The extracted features from these models were fused into one pool to train conventional machine learning classifiers. Peng and Zhang [32] presented a deep learning network based on multiple feature fusion as well as ensemble learning approaches for the diagnosis and treatment of lung diseases. A deep supervised ensemble learning network was used to combine multiple inducers to improve lung lobe segmentation. Tu et al. [33] proposed a general framework for solving online packing problems using deep reinforcement learning hyper-heuristics. They used feature fusion to combine the visual information of real-time packing with distributional information of random parameters of the problem. Tan et al. [34] performed component identification using a deep learning network based on coarse-grained feature fusion. Medjahed et al. [35] fused CNNs trained on different modalities using machine learning (ML) algorithms in the classification phase. Ma et al. [36] proposed a deep dual-side learning ensemble model for Parkinson’s disease diagnosis by analyzing speech data. Their approach employs a weighted fusion mechanism to integrate multiple models. Alzubaidi et al. [37] detected shoulder abnormalities. They trained models using different body part images in the same domain and performed feature fusion with different machine learning classifiers. Agarwal et al. [38] used an approach that combined channel-based fusion and model-based fusion to classify chest x-ray images using ResNet50V3 and InceptionV3 models. Li et al. [39] designed a dual-channel feature fusion network for detecting distal radius fractures. They used Faster region-based CNN (RCNN)and ResNet50 on their channels. The feature fusion method includes an attention mechanism.
While a broader overview of deep learning for X-ray analysis is available in the literature [40], the foundation for the current study is the authors' prior work [1]. That paper details the creation of the dataset used herein and presents a comparative performance analysis of 11 different deep learning models.
The dataset presented in [1] is highly challenging due to two primary factors. Firstly, the overlapping nature of internal components in X-ray images leads to high intra-class variation. Secondly, high inter-class similarity, resulting from the visual resemblance between benign laptop circuits and threat circuits, reduces the distinction between the classes. Collectively, these issues caused the models evaluated in the previous work [1] to be prone to overfitting, a problem exacerbated by the dataset's limited size and high complexity. To overcome this issue and improve classification performance, it was proposed to use the RWN network for the fusion of the feature extraction capabilities of the pre-trained models in a way that is resistant to overfitting. This was anticipated to increase fusion success due to both the single-stage nature of the training and its resistance to overfitting.
Therefore, the motivation for this study is twofold: to address the identified gap in the literature for this specific security application and to overcome the performance limitations of conventional deep learning models that were observed in our prior work [1].

2. Feature Extraction Using Deep Learning

The fundamental difference between deep learning architectures and classical artificial neural networks is their ability to extract features from raw data. This is particularly popular in image processing, analysis, and classification. One significant reason for this popularity is that each pixel can be considered as a feature, leading to millions of features depending on the image resolution. The primary purpose of convolution filters used in image analysis or classification is to extract a feature map. Typically, due to the high number of extracted features, there is a pooling layer following the filter layers. The general scheme of convolution filters is illustrated in Figure 1, and pooling is depicted in Figure 2.
Different architectures have been proposed in the literature for various purposes, such as preventing the vanishing gradient problem by using a different number of convolution filters and pooling, improving the connectivity structure between layers, scalability, reducing computational costs, and achieving higher-performing models. In the scope of this study, ResNet, DarkNet, EfficientNet, DenseNet, MobileNet, InceptionV3, ShuffleNet, and Xception models were trained with the dataset, and features were extracted from these models. After obtaining the features, various machine learning methods such as kNN [41,42,43], SVM [42,43], and random forest [42,43] have been applied in the literature. In the scope of this study, feature vectors from 11 deep learning methods were obtained. An example of extracting and obtaining features after the training of the ResNet18 deep learning architecture is illustrated in Figure 3.
After training, data are passed through the network, and features are extracted from the outputs of intermediate layers, such as convolutional or pooling layers, prior to the final fully connected layer. Table 1 summarizes the models used, the specific layers from which features are extracted, and the corresponding number of extracted features.
Table 1 Facilitates the comparison of different deep learning models based on the layers chosen for feature extraction and the dimensionality of the extracted features. The number of features reflects the complexity and expressive power of each model in capturing information from the input data.

2.1. Late Feature Fusion and Random Weight Network

After extracting N features from one deep learning method and M features from another, two datasets are obtained. These datasets, combined in the forms of (N | M) and (M | N), are subsequently used for training and classification. The overall process of feature extraction, merging, training, and testing is illustrated in Figure 4.
In Figure 4, after feature extraction, the RWN is employed for the training and testing process. For the training of an RWN, the network must have a three-layered structure consisting of an input layer, a hidden layer, and an output layer [44,45,46,47]. The structure of the RWN architecture is shown in Figure 5.
In Figure 5, the features extracted from deep learning networks are provided to the input layer. The weights between the input layer and the hidden layer are randomly initialized. The Moore–Penrose approach is used between the hidden layer and the output layer. Unlike networks trained with iterative algorithms like backpropagation, an RWN calculates the output layer weights analytically in a single step. This non-iterative process results in significantly faster training time. In the current study, the RWN model and training algorithm proposed in [46,47] are used. In a feedforward artificial neural network with N neurons in the hidden layer and an activation function denoted as g(x), the output calculation ( O j ) for M training examples ( x j , t j ) is performed using Equation (1).
O j = i = 1 N β i g w i . x j + b i , j = 1 . . M
where x j = x j 1 , , x j n T R n represents the input vector, t i = t j 1 , , t j m T R m is the target label vector, w i = w i 1 , , w i n T is the weight vector between the input and hidden layers, β i = β i 1 , , β i m T is the weight vector between the hidden and output layers and b i denotes the bias of the i t h hidden neuron. By considering the explanations mentioned above, the model can be written as follows.
H β = T H = g w 1 . x 1 + b 1 g w N . x 1 + b N g w 1 . x M + b 1 g w N . x M + b N β = β 1 T β N T , T = t 1 T t M T
In Equation (2), H is the output matrix of the hidden layer. To train the feedforward artificial neural network, specific weights β i and w i are calculated using a gradient-based learning algorithm as follows:
H w 1 , , w N , b 1 , , b N β T = min w i , b i , β H w 1 , , w N , β T
In the RWN, after the weights between the input layer and the hidden layer are randomly assigned, the weights ( β ) between the hidden layer and the output layer can be calculated as in Equation (4).
β = H + T
where, H + is the Moore–Penrose generalized inverse matrix. In an RWN, if the number of training examples equals the number of hidden neurons, the hidden layer output matrix H becomes a square matrix, and its direct inverse can be computed without the need for the Moore–Penrose approach. However, since the number of training examples is generally greater than the number of neurons in the hidden layer, the Moore–Penrose approach is used to compute the inverse of the matrix. The reason for the fast training process of an RWN is shown in Equation (4), while iterative methods are used to determine the weights given in Equation (3). In the current study, the classification` performance of the RWN was measured using the accuracy value, which represents the ratio of correctly predicted samples by the model.

2.2. Dataset

The study focuses on the development of deep learning methods for the detection of concealed circuits within laptops and emphasizes feature fusion. In this context, there is a need for a dataset to train, test, and extract features from deep learning methods. To meet this requirement, a multi-step process was followed to generate the dataset. This process involved the following:
  • Creating the circuits and/or elements to be concealed;
  • Procuring a variety of laptops;
  • Capturing X-ray images of the circuits after they were hidden inside the laptops.
Readily available Arduino boards were chosen as the example circuits for concealment. Due to the high prices of new laptops and the need for many different laptops, second-hand laptops were purchased from the market. With permission obtained from the Konya Airport Civil Administration Directorate, X-ray images of 60 laptops with different configurations were obtained using X-ray devices at the airport. X-ray images of laptops taken from different angles are provided in Figure 6. In Figure 6, the areas enclosed in red rectangles contain hidden circuits not belonging to the computer motherboard, and the rectangles are drawn manually.
As seen in Figure 6, the image of the laptop's motherboard and the embedded circuit is enclosed in a plastic box (standard practice at airports). Since the plastic transport box is irrelevant for classification, its image was removed from the background using a thresholding technique. This process also eliminated the bright yellow artifacts present in the images. In Figure 7, a clean X-ray image of the laptop is presented.
A total of 6395 X-ray images were captured. Among them, 2545 contain hidden circuits, while 3850 do not. To ensure balance in the data, the number of images without circuits was reduced to 2549, and a total of 5094 X-ray images were used in the experiments. Since the problem is treated as a classification problem, labels for the images are necessary during the training and testing processes with deep learning methods. The 5094 X-ray images were marked as normal or abnormal and stored in different folders. Due to variations in input dimensions in the literature analysis and the deep learning architectures used, the images were resized for each deep learning algorithm to match the input size. Although the width of the conveyor belt of the X-ray machine (number of columns in the image) is expressed as 704 pixels, the length varies due to the continuous flow behavior of the belt (this is not clearly seen in Figure 6 due to its white background). Therefore, as shown in Figure 7, the background and object images were segmented, and a clean image was obtained. This image was then adapted to fit the input of each deep learning method.

3. Experiments

All experiments were conducted on a workstation equipped with an AMD Ryzen 9 5950X 16-Core Processor and 64 GB of RAM. The MATLAB 2021a environment was used for all stages of the study, including feature extraction, feature fusion, and the training and evaluation of the classification models.
The feature extraction process requires deep learning models to first be trained on the dataset. A comprehensive performance analysis of 11 such architectures was previously presented in [1]. The results from that study showed that ShuffleNet achieved the highest test accuracy of 83.55%, followed by the InceptionV3 architecture at 81.31%. In this study, the aim was to achieve higher accuracy in classification using features extracted from these architectures. The number of features extracted from the methods in the [1] study is presented in Table 2.
The features for the datasets obtained by combining the features given in Table 2 are provided in Table 3.
As seen in Table 3, besides using features from different architectures, the feature set from each architecture was concatenated with itself to investigate the effect of this repetition on classification performance. The datasets were used to conduct classification experiments with RWN, employing 10-fold cross-validation in each experiment. To account for the stochastic nature of the RWN, where input-to-hidden layer weights are randomly assigned, we repeated each 10-fold cross-validation experiment 30 times to obtain statistically robust performance measures. Our investigation focused on two key hyperparameters that govern RWN's behavior: the number of hidden layer neurons and the choice of activation function. The investigation began by evaluating the impact of the number of hidden neurons, testing the values as follows: 50, 100, 250, 500, 1000, 2000, and 4585. The value of 4585 was specifically chosen because it matches the number of training samples in each fold of our 10-fold cross-validation. When the number of hidden neurons equals the number of training samples, the hidden layer output matrix (H) becomes a square matrix, allowing its inverse to be calculated directly without requiring the Moore–Penrose pseudo-inverse method. Subsequently, the effect of the activation function, the second key hyperparameter, was investigated. For these experiments, the number of hidden neurons was fixed to the value that yielded the best average test accuracy from the previous stage. The activation functions evaluated include tangent sigmoid, sigmoid, sine, hard limit, triangular basis, and radial basis, all of which are commonly used with RWNs.
Firstly, the features given in Table 2 were reclassified using RWN and compared with the results of deep learning methods. Tangent sigmoid was used as the activation function in RWN, and the comparison results are presented in Table 4.
Table 4 reveals a substantial improvement in test accuracy when features extracted from deep learning models are classified using an RWN. Specifically, while the best-performing standalone deep learning model (ShuffleNet) achieved an accuracy of 83.55%, this figure increased to 94.82% when using an RWN with 250 hidden neurons on the features extracted from the same ShuffleNet model. Additionally, the results indicate a trade-off related to the number of hidden neurons: increasing the neuron count improves training accuracy at the cost of decreasing test accuracy. This trend, which is indicative of overfitting, is illustrated for several model architectures in Figure 8.
A similar analysis using classical machine learning algorithms shows that features extracted from ShuffleNet consistently yield the best performance. Both SVM and KNN demonstrated strong generalization, achieving test accuracies of 93.62% and 94.76%, respectively, on the ShuffleNet features, results that are comparable to those of the RWN. In contrast, while the TREE model achieved high accuracy on the training set, its performance dropped significantly on the test set, clearly emphasizing its tendency to overfit. This comparison highlights the superior generalization capabilities of RWN, SVM, and KNN for this classification task.
A combined evaluation of Table 4 and Figure 8 indicates that increasing the number of hidden neurons leads to overfitting, where the network begins to memorize the training data rather than generalizing from it. The results show that setting the number of hidden neurons to 250 provides the best balance, yielding the highest average test accuracy and overall classification performance among the tested values. Given the dramatic decrease in test performance observed with 4585 neurons, a clear sign of severe overfitting, this value was excluded from subsequent experiments on the combined feature datasets. The results of the experiments using the remaining neuron counts on these combined datasets are presented in Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10.
The extensive feature combinations detailed in Table 3 were designed to systematically investigate the principles of an effective fusion strategy. Our analysis of the subsequent classification results (Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13) revealed several key patterns. Firstly, experiments involving self-combination (e.g., ShuffleNet features fused with themselves) demonstrated no significant performance improvement over using the single feature set with an RWN. This critical finding indicates that merely increasing feature quantity is insufficient; feature diversity is a crucial driver of success. Secondly, the most substantial accuracy gains came from a synergistic fusion of features from the top-performing individual models, ShuffleNet and InceptionV3, where their distinct representational strengths complemented each other to create a more robust and discriminative feature space. This synergy proved more impactful than raw feature dimensionality alone, as this combination outperformed fusions with a higher total feature count. Finally, our tests also confirmed that the order of feature concatenation (e.g., N|M vs. M|N) had a negligible impact on the final classification outcome.
A holistic analysis of Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13 reveals two key trends. First, duplicating the feature set by merging a dataset with itself does not yield a significant improvement in test performance. Second, as previously noted, increasing the number of neurons in the RWN's hidden layer consistently improves training performance, often at the expense of test performance. Furthermore, a critical finding is that fusing features from the individually best-performing deep learning models, notably ShuffleNet and InceptionV3, leads to the highest classification accuracies. This specific combination consistently produced the top results across different classifiers. Specifically, on the fused ShuffleNet-InceptionV3 feature set, several classifiers achieved high training accuracies, with SVM achieving 99.91%, TREE 99.54%, an RWN with 2000 neurons 99.69%, and KNN 97.87%. The highest test accuracy of 97.43% is achieved when the RWN hidden layer neuron count is set to 500 or 1000, and features extracted from Inception and ShuffleNet architectures are combined. As detailed in the overall performance comparison in Table 14, other classifiers like SVM and KNN also achieved high performance on the fused feature sets, though the RWN provided a superior balance of accuracy and efficiency.
Having established the impact of the hidden neuron count, the investigation now shifts to evaluating the influence of the second key hyperparameter: the activation function. To conduct this analysis, the number of hidden neurons was fixed at 250. This value was chosen based on the results in Table 14, as it yielded the highest average test accuracy in the previous experiments. The performances of the different activation functions on the combined datasets are subsequently presented in Table 15, Table 16, Table 17, Table 18 and Table 19. Note that the results for the Tangent Sigmoid function with 250 neurons, which were already presented in Table 7, are not duplicated in this section.
A comparative analysis of the results from Table 7 and Table 15, Table 16, Table 17, Table 18 and Table 19 reveals a clear distinction in the performance of the tested activation functions. The Sigmoid, Tangent Sigmoid, and Hardlimit functions consistently yielded strong and comparable results. In contrast, the Sine, Tribas, and Radbas functions were demonstrably less effective, with their average training and test accuracies remaining below 90%.
The best overall performance was unequivocally achieved using the Sigmoid activation function. On the fused Inception-ShuffleNet feature set, this configuration produced not only the highest training accuracy of 97.74% and a test accuracy of 97.40% but also the highest average training and test accuracies of 93.54% and 92.77%, respectively. A summary of these comparative results is presented in Table 20.
The summary results in Table 20 highlight a clear performance hierarchy among the activation functions. Sigmoid, Tangent Sigmoid, and Hardlimit consistently emerge as the top-performing functions. Conversely, Sine, Tribas, and Radbas demonstrate markedly inferior performance, particularly with respect to their average test accuracies.
Figure 9 presents the time and disk space usage for tests conducted with individual feature sets, while Figure 10 illustrates the same metrics for tests performed with combined feature sets. The reported values are calculated as the averages of the consumption metrics across all tests for each classifier. When evaluating computational efficiency, the RWN demonstrates a strong and balanced time–performance profile. While its training time is longer than that of lazy learners like KNN, it is significantly faster than SVM. More critically, its testing time is remarkably short, outperforming the much slower KNN. SVM exhibits the longest training time of all classifiers but has a faster testing time compared to KNN. KNN, as a lazy learning algorithm, has negligible training time that is limited to loading instances into memory. However, its testing time is significantly longer due to the need to search for nearest neighbors during inference. TREE, on the other hand, is fast in both training and testing but, as demonstrated earlier, is highly prone to overfitting. In terms of disk usage, RWN is the most economical, requiring the least space, whereas SVM is the most demanding. Notably, when moving from individual to combined feature sets, the disk space consumption for KNN, TREE, and SVM increases significantly, while RWN's usage remains low and consistent.
Therefore, when considering the combined metrics of high classification accuracy, minimal disk space requirements, and a favorable balance of training/testing times, the RWN emerges as the most well-rounded and efficient classifier for this application. While SVM and KNN offer high accuracy potential, they demand greater computational and storage resources. TREE achieves a balanced trade-off between time and resource usage, but its classification performance does not match the other classifiers.
Having established the performance of the proposed RWN-based feature fusion framework, the final stage of our analysis compares these results against several state-of-the-art machine learning classifiers. Machine learning advancements have led to a variety of classifiers designed to solve complex problems with varying efficiency. To assess our proposed method, we compared its performance with state-of-the-art classifiers. Among these classifiers, CatBoost is a gradient boosting algorithm designed to handle categorical data effectively while mitigating overfitting [48]. Decision trees use a tree-like structure to model decisions and their possible outcomes. They are widely used for both classification and regression problems [49]. The Gaussian Naïve Bayes algorithm, based on Bayes' theorem, is a probabilistic classifier that calculates the likelihood of different classes. It gained popularity for its effectiveness in classification tasks [50]. Gradient boosting methods enhance model performance by iteratively combining weak learners to create a strong predictive model [51]. KNN is a straightforward yet effective algorithm for classification and regression. It assigns class labels by analyzing the nearest k data points in the feature space [52]. LightGBM is a gradient boosting framework optimized for handling large datasets efficiently through distributed learning [53]. Logistic regression is a statistical method used to predict the probability of a dependent variable belonging to a particular category [54]. Random Forests combine multiple decision trees to tackle complex classification problems, improving accuracy and robustness [55]. The Ridge Classifier is an extension of ridge regression; this algorithm is tailored for classification tasks. It excels by incorporating regularization to address overfitting [56]. SVM is an algorithm that finds the most appropriate hyperplane to separate data into different classes and is widely used for both linear and nonlinear problems [57]. XGBoost is a high-performance gradient boosting framework known for its speed and scalability [58]. In Table 21, we present an analysis of the results, highlighting the strengths and generalization capabilities of the proposed approach.
The comparative results are presented in Table 21. It is crucial to note the experimental setup for this comparison: to provide a robust benchmark, the state-of-the-art classifiers were trained on the best-performing single feature set (ShuffleNet). Our proposed method, by contrast, was trained on the fused ShuffleNet-InceptionV3 feature set to specifically demonstrate the benefit of feature fusion.
The analysis clearly demonstrates the superiority of the proposed method. The RWN-based fusion approach not only achieves the highest average test accuracy of 97.43% but also exhibits the best generalization capability. This is evidenced by the minimal gap between its training and test accuracies, especially when compared to models like LightGBM and XGBoost, which, despite achieving perfect training scores, show a significant performance drop on the test set, indicating severe overfitting.

4. Results and Discussion

This study systematically investigated the performance of a novel feature fusion framework centered on a Random Weight Network (RWN) classifier. The findings directly address the core research questions posed in the Introduction, demonstrating a clear pathway to overcoming the limitations of conventional deep learning models in this challenging security domain.
The investigation first addressed the performance implications of substituting a standard deep learning classifier with an RWN. The results unequivocally demonstrate a significant performance uplift. For instance, on the features extracted from the best-performing standalone model, ShuffleNet, the test accuracy increased dramatically from 83.55% to 94.82% when an RWN with 250 hidden neurons was employed. This finding confirms that by decoupling feature extraction from classification, the inherent performance limitations of conventional models, namely low accuracy and a high propensity for overfitting on complex X-ray data, can be substantially mitigated.
Building upon this, the study validated its primary hypothesis regarding the efficacy of feature fusion. By combining features from different high-performing architectures, notably ShuffleNet and InceptionV3, the framework achieved a state-of-the-art test accuracy of 97.44%. This result provides a definitive affirmative answer to whether multi-model fusion can significantly enhance classification accuracy, clearly outperforming both standalone models and the RWN applied to single feature sets. This highlights that data diversity, achieved through fusing varied feature representations, is a key driver of performance. In contrast, simply duplicating an existing feature set by merging it with itself yields no significant improvement, reinforcing that the richness of the feature pool is what matters.
The performance of the proposed framework was also found to be critically dependent on the careful tuning of RWN’s hyperparameters. Addressing the impact of hidden layer size, the study revealed a clear trade-off: an excessive number of neurons led to overfitting, while an insufficient number resulted in ineffective learning. Optimal generalization was achieved through a balance, with 250 neurons providing the best average test accuracy across many scenarios, and 500 or 1000 neurons yielding the peak accuracy on the best fused dataset. Similarly, the choice of activation function proved significant. Sigmoid, Tangent Sigmoid, and Hardlimit functions consistently delivered superior and robust performance, with Sigmoid ultimately achieving the best overall results. Conversely, other implementation details, such as the order of feature concatenation (N|M vs. M|N), were found to have a negligible impact on the outcome.
In summary, this study confirms that a modular approach, which involves decoupling feature extraction, employing multi-model feature fusion, and utilizing a well-tuned RWN, is a highly effective strategy. This framework successfully answers the initial research challenges, demonstrating a clear pathway from the 83.55% accuracy of standalone models to the 97.44% achieved through the proposed methodology, thereby establishing a new performance benchmark in this security domain.

5. Conclusion and Future Works

This study successfully addressed the challenges of low accuracy and overfitting in deep learning-based X-ray image classification by proposing a novel framework based on feature fusion and a Random Weight Network (RWN) classifier. We demonstrated that by fusing features from diverse, high-performing deep learning architectures, specifically ShuffleNet and InceptionV3, and using a well-tuned RWN, the classification performance can be dramatically improved. The core contribution of this work is the significant increase in test accuracy, from a baseline of 83.55% for the best standalone model to a state-of-the-art 97.44% with the proposed fusion method. Our analysis confirmed that the success of this framework depends not only on the diversity of the fused features but also on the careful tuning of the RWN's hyperparameters, namely the number of hidden neurons and the choice of activation function. This modular approach provides a robust and computationally efficient alternative to end-to-end deep learning systems for this critical security task.
For future work, we plan to expand on these findings by investigating a broader range of feature combination methods and testing the performance of other advanced classifiers. Furthermore, we plan to explore transformer-based feature extraction techniques and deep fusion techniques. In this approach, instead of combining the features after they are fully extracted, the integration would happen at earlier or intermediate stages within a single, unified neural network.

Author Contributions

Conceptualization: M.S.K. and E.E.; Methodology, M.S.K., G.S., E.E. and M.Y.; Software, M.S.K. and G.S.; Validation, M.S.K., G.S., E.E. and X.W.; Formal Analysis, M.S.K. and X.W.; Investigation, G.S. and M.Y.; Resources, M.S.K. and E.E.; Data curation, G.S. and M.Y.; Writing—original draft preparation, M.S.K., G.S. and E.E.; Writing—review and editing, M.S.K., E.E., G.S., M.Y. and X.W.; Visualization, G.S. and M.Y.; Supervision, M.S.K. and X.W.; Project administration, M.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data can be available at http://mskiran.ktun.edu.tr/ecd (accessed on 13 August 2025).

Acknowledgments

This study has been supported by the Scientific and Technological Research Council of Türkiye with grant number [122E024]. The authors (M.S. Kiran, E. Esme, G. Seyfi, and M. Yilmaz) would like to thank the council for their institutional support. The author, Xizhao Wang, would like to acknowledge the Stable Support Project of Shenzhen City (No. 20231122124602001).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Seyfi, G.; Yilmaz, M.; Esme, E.; Kiran, M.S. X-Ray Image Analysis for Explosive Circuit Detection using Deep Learning Algorithms. Appl. Soft Comput. 2023, 151, 11133. [Google Scholar] [CrossRef]
  2. Wikipedia. X-Ray. Available online: https://en.wikipedia.org/wiki/X-ray (accessed on 26 August 2022).
  3. Singh, S.; Singh, M. Explosives detection systems (EDS) for aviation security. Signal Process. 2003, 83, 31–55. [Google Scholar] [CrossRef]
  4. Mery, D.; Riffo, V.; Zuccar, I.; Pieringer, C. Automated X-ray object recognition using an efficient search algorithm in multiple views. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Portland, OR, USA, 23–28 June 2013; pp. 368–374. [Google Scholar] [CrossRef]
  5. Schmidt-Hackenberg, L.; Yousefi, M.R.; Breuel, T.M. Visual cortex inspired features for object detection in X-ray images. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; pp. 2573–2576. [Google Scholar] [CrossRef]
  6. Akçay, S.; Kundegorski, M.E.; Devereux, M.; Breckon, T.P. Transfer learning using convolutional neural networks for object classification within x-ray baggage security imagery. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phonix, AZ, USA, 25–28 September 2016; pp. 1057–1061. [Google Scholar] [CrossRef]
  7. Jaccard, N.; Rogers, T.W.; Morton, E.J.; Griffin, L.D. Tackling the X-ray cargo inspection challenge using machine learning. In Proceedings of the Anomaly Detection and Imaging with X-Rays (ADIX), Baltimore, MD, USA, 19–20 April 2016; p. 98470N. [Google Scholar] [CrossRef]
  8. Mery, D.; Svec, E.; Arias, M.; Riffo, V.; Saavedra, J.M.; Banerjee, S. Modern computer vision techniques for x-ray testing in baggage inspection. IEEE Trans. Syst. Man Cybern. Syst. 2016, 47, 682–692. [Google Scholar] [CrossRef]
  9. Jaccard, N.; Rogers, T.W.; Morton, E.J.; Griffin, L.D. Detection of concealed cars in complex cargo X-ray imagery using deep learning. J. X-Ray Sci. Technol. 2017, 25, 323–339. [Google Scholar] [CrossRef]
  10. Rogers, T.W.; Jaccard, N.; Griffin, L.D. A deep learning framework for the automated inspection of complex dual-energy X-ray cargo imagery. In Proceedings of the Anomaly Detection and Imaging with X-Rays (ADIX) II, Anaheim, CA, USA, 12–13 April 2017; p. 101870L. [Google Scholar] [CrossRef]
  11. Caldwell, M.; Ransley, M.; Rogers, T.W.; Griffin, L.D. Transferring x-ray based automated threat detection between scanners with different energies and resolution. In Proceedings of the Counterterrorism, Crime Fighting, Forensics, and Surveillance Technologies, Warsaw, Poland, 11–14 September 2017; p. 104410F. [Google Scholar] [CrossRef]
  12. Morris, T.; Chien, T.; Goodman, E. Convolutional neural networks for automatic threat detection in security X-Ray images. In Proceedings of the 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 285–292. [Google Scholar] [CrossRef]
  13. Akcay, S.; Breckon, T.P. An evaluation of region based object detection strategies within X-ray baggage security imagery. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1337–1341. [Google Scholar] [CrossRef]
  14. Akcay, S.; Kundegorski, M.E.; Willcocks, C.G.; Breckon, T.P. Using deep convolutional neural network architectures for object classification and detection within x-ray baggage security imagery. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2203–2215. [Google Scholar] [CrossRef]
  15. Petrozziello, A.; Jordanov, I. Automated deep learning for threat detection in luggage from X-ray images. In Proceedings of the International Symposium on Experimental Algorithms, Kalamata, Greece, 24–29 June 2019; pp. 505–512. [Google Scholar] [CrossRef]
  16. Cheng, Q.; Lan, T.; Cai, Z.; Li, J. X-YOLO: An Efficient Detection Network of Dangerous Objects in X-ray Baggage Images. IEEE Signal Process. Lett. 2024, 31, 2270–2274. [Google Scholar] [CrossRef]
  17. Wu, J.; Xu, X. EslaXDET: A new X-ray baggage security detection framework based on self-supervised vision transformers. Eng. Appl. Artif. Intell. 2024, 127, 107440. [Google Scholar] [CrossRef]
  18. Wang, S.; Wang, S.; Xiao, Z. Feature extraction method with efficient edge information enhancement for detecting dangerous objects in security x-ray images. In Proceedings of the International Conference on Computer Graphics, Artificial Intelligence, and Data Processing (ICCAID 2024), Nanchang, China, 13–15 December 2025; pp. 95–106. [Google Scholar] [CrossRef]
  19. Yang, J.; Zhao, Z.; Zhang, H.; Shi, Y. Data augmentation for X-ray prohibited item images using generative adversarial networks. IEEE Access 2019, 7, 28894–28902. [Google Scholar] [CrossRef]
  20. Kaminetzky, A.; Mery, D. In-depth analysis of automated baggage inspection using simulated X-ray images of 3D models. Neural Comput. Appl. 2024, 36, 18761–18780. [Google Scholar] [CrossRef]
  21. Caldwell, M.; Griffin, L.D. Limits on transfer learning from photographic image data to X-ray threat detection. J. X-Ray Sci. Technol. 2019, 27, 1007–1020. [Google Scholar] [CrossRef]
  22. Benedykciuk, E.; Denkowski, M.; Dmitruk, K. Material classification in X-ray images based on multi-scale CNN. Signal Image Video Process. 2021, 15, 1285–1293. [Google Scholar] [CrossRef]
  23. Babalik, A.; Babadag, A. A binary sparrow search algorithm for feature selection on classification of X-ray security images. Appl. Soft Comput. 2024, 158, 111546. [Google Scholar] [CrossRef]
  24. Ayantayo, A.; Kaur, A.; Kour, A.; Schmoor, X.; Shah, F.; Vickers, I.; Kearney, P.; Abdelsamea, M.M. Network intrusion detection using feature fusion with deep learning. J. Big Data 2023, 10, 167. [Google Scholar] [CrossRef]
  25. Zhang, K.; Ting, H.-N.; Choo, Y.-M. Baby cry recognition by BCRNet using transfer learning and deep feature fusion. IEEE Access 2023, 11, 126251–126262. [Google Scholar] [CrossRef]
  26. Wu, X.; Shi, H.; Zhu, H. Fault Diagnosis for Rolling Bearings Based on Multiscale Feature Fusion Deep Residual Networks. Electronics 2023, 12, 768. [Google Scholar] [CrossRef]
  27. Liu, H.; Hou, C.-J.; Tang, J.-L.; Sun, L.-T.; Lu, K.-F.; Liu, Y.; Du, P. Deep learning and ultrasound feature fusion model predicts the malignancy of complex cystic and solid breast nodules with color Doppler images. Sci. Rep. 2023, 13, 10500. [Google Scholar] [CrossRef]
  28. Patil, S.; Kirange, D. An Optimized Deep Learning Model with Feature Fusion for Brain Tumor Detection. Int. J. Next-Gener. Comput. 2023, 14. [Google Scholar] [CrossRef]
  29. Gill, H.S.; Murugesan, G.; Mehbodniya, A.; Sajja, G.S.; Gupta, G.; Bhatt, A. Fruit type classification using deep learning and feature fusion. Comput. Electron. Agric. 2023, 211, 107990. [Google Scholar] [CrossRef]
  30. Deng, T.; Wang, H.; He, D.; Xiong, N.; Liang, W.; Wang, J. Multi-Dimensional Fusion Deep Learning for Side Channel Analysis. Electronics 2023, 12, 4728. [Google Scholar] [CrossRef]
  31. Al-Timemy, A.H.; Alzubaidi, L.; Mosa, Z.M.; Abdelmotaal, H.; Ghaeb, N.H.; Lavric, A.; Hazarbassanov, R.M.; Takahashi, H.; Gu, Y.; Yousefi, S. A Deep Feature Fusion of Improved Suspected Keratoconus Detection with Deep Learning. Diagnostics 2023, 13, 1689. [Google Scholar] [CrossRef]
  32. Peng, Y.; Zhang, J. Lung lobe segmentation in computed tomography images based on multi-feature fusion and ensemble learning framework. Int. J. Imaging Syst. Technol. 2023, 33, 2088–2099. [Google Scholar] [CrossRef]
  33. Tu, C.; Bai, R.; Aickelin, U.; Zhang, Y.; Du, H. A deep reinforcement learning hyper-heuristic with feature fusion for online packing problems. Expert Syst. Appl. 2023, 230, 120568. [Google Scholar] [CrossRef]
  34. Tan, J.; Wan, J.; Xia, D. Automobile Component Recognition Based on Deep Learning Network with Coarse-Fine-Grained Feature Fusion. Int. J. Intell. Syst. 2023, 2023, 1903292. [Google Scholar] [CrossRef]
  35. Medjahed, C.; Mezzoudj, F.; Rahmoun, A.; Charrier, C. Identification based on feature fusion of multimodal biometrics and deep learning. Int. J. Biom. 2023, 15, 521–538. [Google Scholar] [CrossRef]
  36. Ma, J.; Zhang, Y.; Li, Y.; Zhou, L.; Qin, L.; Zeng, Y.; Wang, P.; Lei, Y. Deep dual-side learning ensemble model for Parkinson speech recognition. Biomed. Signal Process. Control 2021, 69, 102849. [Google Scholar] [CrossRef]
  37. Alzubaidi, L.; Salhi, A.; A. Fadhel, M.; Bai, J.; Hollman, F.; Italia, K.; Pareyon, R.; Albahri, A.; Ouyang, C.; Santamaría, J. Trustworthy deep learning framework for the detection of abnormalities in X-ray shoulder images. PLoS ONE 2024, 19, e0299545. [Google Scholar] [CrossRef]
  38. Agarwal, S.; Arya, K.; Meena, Y.K. Multifusionnet: Multilayer multimodal fusion of deep neural networks for chest x-ray image classification. arXiv 2024, arXiv:2401.00728. [Google Scholar] [CrossRef]
  39. Li, J.; Shan, H.-J.; Yu, X.-W. Fracture detection of distal radius using deep-learning-based dual-channel feature fusion algorithm. Chin. J. Traumatol. 2025, 1–13. [Google Scholar] [CrossRef]
  40. Seyfi, G.; Esme, E.; Yilmaz, M.; Kiran, M.S. A literature review on deep learning algorithms for analysis of X-ray images. Int. J. Mach. Learn. Cybern. 2024, 15, 1165–1181. [Google Scholar] [CrossRef]
  41. Sani, S.; Wiratunga, N.; Massie, S. Learning deep features for kNN-based human activity recognition. In Proceedings of the ICCBR 2017 Workshops, Trondheim, Norway, 26–28 June 2017. [Google Scholar] [CrossRef]
  42. Singh, J.; Thakur, D.; Ali, F.; Gera, T.; Kwak, K.S. Deep feature extraction and classification of android malware images. Sensors 2020, 20, 7013. [Google Scholar] [CrossRef]
  43. Benyahia, S.; Meftah, B.; Lézoray, O. Multi-features extraction based on deep learning for skin lesion classification. Tissue Cell 2022, 74, 101701. [Google Scholar] [CrossRef]
  44. Pao, Y.-H.; Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 1992, 25, 76–79. [Google Scholar] [CrossRef]
  45. Schmidt, W.F.; Kraaijveld, M.A.; Duin, R.P. Feed forward neural networks with random weights. In Proceedings of the International Conference on Pattern Recognition, The Hague, The Netherlands, 30 August–3 September 1992; p. 1. [Google Scholar] [CrossRef]
  46. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the IEEE international Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar] [CrossRef]
  47. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  48. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar] [CrossRef]
  49. Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar] [CrossRef]
  50. Mitchell, T.M.; Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997; Volume 1. [Google Scholar]
  51. Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobot. 2013, 7, 21. [Google Scholar] [CrossRef]
  52. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  53. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  54. Cox, D.R. The regression analysis of binary sequences. J. R. Stat. Soc. Ser. B Stat. Methodol. 1958, 20, 215–232. [Google Scholar] [CrossRef]
  55. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  56. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  57. Cortes, C.; Vapnik, V. Support-Vector Networks. In Machine Learning; Springer: Berlin/Heidelberg, Germany, 1995. [Google Scholar]
  58. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
Figure 1. The extraction of feature maps through filtering on an image.
Figure 1. The extraction of feature maps through filtering on an image.
Applsci 15 09053 g001
Figure 2. 2 × 2 Maximum pooling.
Figure 2. 2 × 2 Maximum pooling.
Applsci 15 09053 g002
Figure 3. ResNet model and the layer where features are extracted from the ResNet architecture.
Figure 3. ResNet model and the layer where features are extracted from the ResNet architecture.
Applsci 15 09053 g003
Figure 4. The general diagram of feature extraction, fusion and classification in the study.
Figure 4. The general diagram of feature extraction, fusion and classification in the study.
Applsci 15 09053 g004
Figure 5. A general architecture of Random Weight Networks.
Figure 5. A general architecture of Random Weight Networks.
Applsci 15 09053 g005
Figure 6. X-ray images of laptops obtained from different angles.
Figure 6. X-ray images of laptops obtained from different angles.
Applsci 15 09053 g006
Figure 7. Cleaned and segmented X-ray image.
Figure 7. Cleaned and segmented X-ray image.
Applsci 15 09053 g007
Figure 8. The effect of the number of neurons in the RWN hidden layer on training and test accuracy.
Figure 8. The effect of the number of neurons in the RWN hidden layer on training and test accuracy.
Applsci 15 09053 g008aApplsci 15 09053 g008b
Figure 9. Time consumption and disk usage for individual feature sets.
Figure 9. Time consumption and disk usage for individual feature sets.
Applsci 15 09053 g009
Figure 10. Time consumption and disk usage for combined feature sets.
Figure 10. Time consumption and disk usage for combined feature sets.
Applsci 15 09053 g010
Table 1. Layers from which features are extracted and the number of extracted features from deep learning models.
Table 1. Layers from which features are extracted and the number of extracted features from deep learning models.
ModelsFeature-Extracted Layer
EfficientNetglobavgpool
ResNet18pool5
ResNet50avgpool
ResNet101pool5
DarkNet19conv19
DarkNet53avg1
MobileNetV2global_average_pooling
ShuffleNetnode_200
InceptionV3avg_pool
DenseNet201avg_pool
Xceptionavg_pool
Table 2. The number of extracted features using deep learning models.
Table 2. The number of extracted features using deep learning models.
ModelsNumber of Extracted Features
EfficientNet1280
ResNet18512
ResNet502048
ResNet1012048
DarkNet19128
DarkNet531024
MobileNetV21280
ShuffleNet544
InceptionV32048
DenseNet2011920
Xception2048
Table 3. The number of combined features obtained from deep learning models.
Table 3. The number of combined features obtained from deep learning models.
ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
EfficientNet25601792332833281408230425601824332832003328
ResNet181792102425602560640153617921056256024322560
ResNet5033282560409640962176307233282592409639684096
ResNet10133282560409640962176307233282592409639684096
DarkNet1914086402176217625611521408672217620482176
DarkNet5323041536307230721152204823041568307229443072
MobileNetV225601792332833281408230425601824332832003328
ShuffleNet1824105625922592672156818241088259224642592
InceptionV333282560409640962176307233282592409639684096
DenseNet20132002432396839682048294432002464396838403968
Xception33282560409640962176307233282592409639684096
Table 4. The comparison of deep learning models and RWN with different numbers of neurons in the hidden layer on the dataset.
Table 4. The comparison of deep learning models and RWN with different numbers of neurons in the hidden layer on the dataset.
ModelsModels’ ResultRWN
(50)
RWN (100)RWN (250)RWN (500)RWN (1000)RWN (2000)RWN (4585)SVMTREEKNN
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.98440.59130.86560.86150.86850.86020.87650.85730.88980.85180.91420.84110.95950.81501.00000.56880.86260.86240.96990.80130.87810.8528
ResNet180.94690.65880.87080.86960.87600.87250.88320.87430.89220.87240.91050.86290.95110.82441.00000.54810.88980.87870.97450.81290.88850.8657
ResNet500.95000.62580.85990.85620.86640.86190.87380.86250.88340.85850.90070.84660.93780.80081.00000.54870.88680.87510.97780.79720.87860.8557
ResNet1010.93440.61010.84980.84820.85490.85160.86110.84910.86840.84560.88830.82860.93310.77841.00000.52530.85940.85730.97350.76830.86110.8341
DarkNet190.89690.78940.88090.87790.88540.88160.89340.88210.90390.88030.92420.87500.96630.84521.00000.57400.89470.88160.97230.83530.89880.8750
DarkNet530.70940.66980.75480.75090.76470.75710.78010.76010.79870.75720.83360.74450.89730.70720.99920.56540.77030.76350.96780.68370.80760.7479
MobileNetV20.95940.60230.84670.84460.84870.84410.85300.84070.86290.83500.88380.82190.93030.77841.00000.52730.84740.84450.97520.77800.86200.8276
ShuffleNet0.99060.83550.94320.94240.94850.94650.95370.94820.95970.94770.97000.94670.99000.93931.00000.57450.98050.93620.99020.91240.95620.9476
InceptionV30.92190.81310.92900.92760.93190.92830.93900.93100.94740.93040.96040.92560.98310.90291.00000.56690.93260.93130.98530.88910.93750.9238
DenseNet2010.88130.77720.90900.90720.91290.90940.92080.91110.93150.91070.95040.90670.98140.88491.00000.61130.86260.86240.96990.80130.87810.8528
Xception0.98440.71260.91520.91260.91780.91260.92400.91120.93350.90720.95000.89960.97800.87391.00000.57330.88980.87870.97450.81290.88850.8657
Table 5. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 50).
Table 5. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 50).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.86590.86180.92680.92490.92050.91850.92000.91830.91760.91520.88230.87960.92500.92380.96250.96090.96350.96240.94330.94180.95260.9513
ResNet180.92680.92560.87030.86840.90380.90090.90230.90060.90890.90660.86760.86450.90170.89940.95200.95060.94190.94040.92410.92180.94050.9385
ResNet500.91980.91870.90330.90050.85780.85550.89890.89580.90240.89990.85540.85180.90490.90280.95200.95000.94090.93880.92330.92100.93270.9303
ResNet1010.91960.91800.90200.89970.89860.89610.84840.84580.89660.89480.85260.84950.90140.89850.94770.94590.94250.94090.92490.92330.93440.9332
DarkNet190.91710.91490.90830.90630.90270.89980.89600.89370.88080.87820.87640.87400.90460.90200.94050.93900.91250.91040.90270.90020.92080.9183
DarkNet530.88290.88140.86770.86420.85440.85230.85250.84920.87670.87410.75140.74690.85920.85760.93320.93130.91790.91600.89220.89060.91040.9084
MobileNetV20.92490.92380.90190.89970.90480.90230.90160.89950.90460.90150.85920.85740.84630.84370.95490.95360.94900.94760.93280.93170.94110.9400
ShuffleNet0.96200.96130.95170.94990.95240.95150.94800.94670.94070.93880.93330.93120.95460.95350.94260.94100.96340.96210.95660.95440.96430.9629
InceptionV30.96340.96220.94200.93970.94100.93920.94240.94110.91170.90940.91740.91520.94930.94750.96340.96220.92850.92650.95360.95270.96040.9592
DenseNet2010.94320.94200.92420.92200.92350.92120.92560.92310.90260.90010.89120.88860.93290.93110.95740.95610.95360.95200.90890.90670.95160.9504
Xception0.95280.95170.94030.93870.93210.93060.93440.93220.92150.91920.91070.90910.94110.94010.96410.96300.96010.95890.95170.95000.91500.9119
Table 6. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 100).
Table 6. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 100).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.86920.85960.93270.92930.92600.92260.92470.92170.92760.92370.89070.88680.92790.92560.96820.96580.96570.96390.94610.94280.95430.9518
ResNet180.93290.93000.87520.87170.91350.90770.91140.90810.92020.91650.88080.87640.90860.90510.96050.95780.94960.94770.93170.92840.94560.9423
ResNet500.92620.92320.91340.90860.86530.86090.90930.90470.91540.91110.87070.86410.91380.90940.96050.95840.94800.94490.93240.92890.93830.9347
ResNet1010.92460.92130.91180.90760.90940.90430.85390.84900.90940.90490.86670.86120.90880.90500.95670.95460.94890.94650.93270.92960.94000.9364
DarkNet190.92690.92270.92060.91660.91550.91090.90910.90390.88550.88090.88390.87920.91590.91080.95330.95080.92500.92170.91370.90910.93200.9276
DarkNet530.89010.88660.88080.87640.87070.86460.86650.86160.88390.87930.76170.75480.87210.86700.94470.94190.92860.92560.90310.89920.91820.9145
MobileNetV20.92790.92490.90870.90490.91370.90950.90900.90480.91600.91200.87220.86760.84870.84380.96190.95960.95320.95120.93720.93540.94450.9416
ShuffleNet0.96830.96640.96030.95790.96060.95850.95650.95410.95350.95050.94460.94180.96230.96020.94850.94600.97050.96850.96290.96100.97020.9680
InceptionV30.96580.96400.94960.94620.94810.94590.94930.94690.92510.92050.92910.92510.95310.95060.97060.96810.93130.92830.95700.95470.96290.9605
DenseNet2010.94620.94270.93150.92770.93220.92930.93240.92930.91360.90870.90270.89950.93730.93420.96280.96040.95700.95460.91260.90870.95490.9525
Xception0.95430.95200.94580.94300.93820.93510.94010.93640.93210.92800.91810.91470.94460.94150.97020.96800.96280.96080.95500.95250.91780.9113
Table 7. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 250).
Table 7. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 250).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.87800.85590.93960.93180.93250.92480.92990.92260.93780.92890.89840.89010.93250.92490.97360.96940.96870.96390.95160.94380.95670.9510
ResNet180.93950.93190.88210.87350.92350.91340.92180.91110.93150.92210.89310.88230.91630.90690.96730.96200.95700.94990.93920.93170.95140.9454
ResNet500.93240.92490.92380.91270.87290.86190.91940.90880.92870.91890.88430.87160.92200.91240.96720.96250.95430.94750.94060.93340.94380.9365
ResNet1010.93010.92180.92140.91140.91970.90850.86040.84950.92400.91270.87870.86800.91720.90790.96400.95890.95600.95000.94060.93330.94630.9383
DarkNet190.93760.92900.93170.92150.92880.91880.92390.91320.89310.88220.89440.88190.92810.91740.96410.95690.93880.92960.92570.91560.94260.9346
DarkNet530.89860.88980.89280.88260.88440.87240.87870.86740.89420.88310.77710.75790.88420.87330.95370.94850.93790.93160.91380.90540.92510.9170
MobileNetV20.93220.92510.91640.90730.92200.91240.91740.90790.92790.91700.88450.87370.85330.84060.96800.96380.95800.95190.94320.93720.94900.9418
ShuffleNet0.97370.96940.96730.96240.96710.96210.96420.95880.96400.95710.95370.94770.96800.96340.95360.94780.97670.97240.96850.96410.97500.9708
InceptionV30.96870.96470.95700.95020.95440.94840.95590.94950.93870.92930.93810.93150.95800.95170.97680.97250.93850.92900.96180.95590.96620.9601
DenseNet2010.95150.94420.93930.93070.94070.93210.94080.93310.92570.91590.91390.90590.94320.93700.96860.96400.96180.95590.92060.90970.95990.9539
Xception0.95680.94940.95150.94460.94390.93600.94630.93870.94220.93400.92530.91750.94900.94110.97510.97050.96600.96040.95990.95390.92390.9100
Table 8. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 500).
Table 8. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 500).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.89230.85120.94610.93190.93890.92300.93620.92130.94770.92920.90620.88910.93850.92330.97820.97090.97320.96390.95980.94300.96150.9481
ResNet180.94610.93190.89100.87110.93300.91260.93030.91130.94220.92360.90390.88230.92390.90540.97260.96340.96340.95030.94740.93200.95760.9444
ResNet500.93910.92250.93290.91260.88170.85810.92900.90810.94010.91980.89500.87040.92980.91130.97250.96350.96050.94650.94780.93230.94950.9346
ResNet1010.93630.92110.93060.91220.92890.90720.86810.84410.93640.91430.88880.86630.92490.90610.97050.96090.96190.94950.94870.93400.95300.9365
DarkNet190.94790.92990.94210.92460.94020.92000.93620.91540.90370.88120.90570.88130.93860.91940.97130.95920.94960.93230.93670.91700.95160.9350
DarkNet530.90620.88800.90370.88320.89510.87130.88840.86600.90560.88240.79330.75530.89450.87290.96020.95050.94530.93200.92260.90640.93190.9155
MobileNetV20.93870.92390.92400.90510.92980.91110.92510.90630.93840.91920.89490.87270.86280.83540.97310.96460.96350.95150.95020.93670.95530.9394
ShuffleNet0.97810.97110.97270.96410.97260.96410.97050.96070.97140.95960.96020.95000.97320.96430.95950.94810.98160.97430.97390.96530.97950.9719
InceptionV30.97330.96320.96360.95090.96060.94690.96180.94940.94950.93270.94520.93310.96360.95090.98160.97440.94690.92990.96800.95600.97070.9593
DenseNet2010.95960.94400.94720.93230.94790.93190.94870.93440.93670.91750.92240.90630.95020.93790.97400.96550.96790.95620.93130.91040.96580.9546
Xception0.96130.94860.95760.94450.94970.93460.95310.93690.95210.93520.93200.91480.95530.93970.97940.97180.97080.95910.96580.95430.93310.9068
Table 9. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 1000).
Table 9. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 1000).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.91830.84100.95770.92780.95030.91780.94770.91670.96290.92870.92010.87670.95190.91810.98470.97060.98220.96010.97390.94100.97180.9433
ResNet180.95770.92710.90810.86150.94780.90510.94520.90510.95890.92120.92100.87460.93810.89850.98030.96270.97290.94850.96040.92850.96840.9397
ResNet500.95040.91640.94760.90700.89840.84450.94430.90120.95760.91830.91270.86140.94300.90350.98050.96340.97070.94230.95920.92660.95970.9265
ResNet1010.94790.91370.94520.90540.94450.90010.88680.82540.95440.91290.90610.85440.93910.89900.97960.96030.97140.94680.96180.93100.96520.9320
DarkNet190.96310.92810.95870.92160.95750.91660.95440.91290.92460.87340.92680.87580.95540.91600.98170.95990.96490.93160.95470.91380.96700.9341
DarkNet530.92000.87680.92110.87600.91270.86010.90620.85350.92680.87630.82400.74090.91120.86220.96960.94930.95550.92850.93680.90050.94340.9091
MobileNetV20.95180.91850.93790.89860.94300.90400.93920.89870.95530.91620.91140.86200.88320.82030.98120.96350.97290.94810.96230.93440.96630.9334
ShuffleNet0.98480.97080.98020.96280.98060.96350.97960.96070.98180.95880.96960.94890.98120.96360.97020.94740.98810.97430.98250.96600.98620.9714
InceptionV30.98220.96120.97300.94870.97070.94280.97150.94700.96490.93030.95560.92790.97280.94860.98790.97440.96070.92250.97900.95500.98020.9540
DenseNet2010.97390.94110.96040.92880.95930.92680.96170.93050.95450.91430.93720.89940.96220.93430.98260.96600.97900.95530.95090.90440.97670.9507
Xception0.97180.94220.96840.94030.95970.92710.96500.93290.96700.93310.94340.90740.96640.93330.98620.97120.98010.95380.97670.95150.94980.8948
Table 10. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 2000).
Table 10. The performance of RWN on the combined dataset (The number of neurons in the hidden layer was set to 2000).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.96190.81470.97980.90740.97300.89100.97320.88880.98670.91410.94830.83430.97680.89530.99640.96830.99510.95160.99340.93060.98980.9218
ResNet180.97990.90780.94710.82330.97230.88150.97350.88160.98590.90600.95580.84210.96680.87120.99360.95840.98890.93550.98300.91210.98600.9254
ResNet500.97300.89030.97250.88060.93280.80210.97140.87220.98590.89960.94650.82060.96870.87370.99310.95870.98670.92640.98020.90480.97930.9034
ResNet1010.97310.88720.97340.88210.97130.87250.93030.77680.98370.89360.94360.80800.96770.86840.99340.95580.98900.93190.98380.91490.98480.9141
DarkNet190.98660.91470.98600.90580.98590.90240.98380.89460.96670.84550.96810.84960.98450.89910.99540.95440.98820.91870.98390.89780.99000.9222
DarkNet530.94850.83570.95570.84220.94660.82160.94340.80770.96820.84760.88240.70820.94400.81950.98750.94050.97470.90290.96500.86590.96610.8764
MobileNetV20.97650.89620.96690.87270.96880.87320.96780.87040.98450.89740.94400.81990.92800.77560.99470.95820.98880.93420.98270.91770.98610.9133
ShuffleNet0.99640.96840.99360.95760.99310.95760.99340.95520.99530.95390.98740.94120.99470.95890.99010.93930.99690.97160.99580.96320.99620.9684
InceptionV30.99510.95200.98900.93670.98670.92490.98900.93210.98820.91870.97470.90270.98890.93420.99680.97230.98330.90010.99440.94720.99390.9399
DenseNet2010.99330.93070.98290.91200.98040.90410.98380.91380.98380.89840.96480.86490.98300.91830.99580.96360.99450.94830.98240.88340.99330.9385
Xception0.98980.92500.98610.92490.97920.90380.98460.91370.99010.92250.96590.87680.98610.91300.99620.96870.99390.93900.99310.93910.97780.8677
Table 11. The performance of SVM on the combined dataset.
Table 11. The performance of SVM on the combined dataset.
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.86290.86260.94610.93580.94080.93010.92850.92480.94580.93560.89890.89580.93040.92680.99620.95920.96770.96580.95210.94620.95400.9521
ResNet180.94610.93580.89300.87910.94290.92230.93560.92010.95100.93230.91360.90070.92490.91300.99620.95190.96540.95680.95240.94050.95760.9474
ResNet500.94080.93010.94290.92230.89140.87630.93250.91790.95080.92990.90530.88910.93350.91950.99660.95490.96670.95660.95390.94030.94890.9403
ResNet1010.92850.92480.93560.92010.93250.91790.86160.85570.93990.92780.88310.87510.91750.91010.99200.94900.96010.95640.95040.94250.94640.9431
DarkNet190.94580.93560.95100.93230.95080.92970.93980.92800.89480.88440.90560.89380.93830.92680.99760.94990.95460.94390.95000.93500.95230.9419
DarkNet530.89890.89580.91360.90070.90530.88910.88310.87510.90550.89380.77190.76310.89060.88380.98510.93880.94270.93910.92530.91910.92450.9201
MobileNetV20.93040.92680.92490.91300.93350.91950.91750.91010.93820.92680.89060.88380.84750.84300.99570.95390.96030.95660.95020.94440.94790.9423
ShuffleNet0.99620.95920.99620.95190.99660.95490.99200.94900.99760.94990.98510.93880.99570.95390.98400.93070.99910.96680.99670.95840.99790.9617
InceptionV30.96770.96580.96540.95680.96670.95660.96010.95640.95460.94390.94270.93910.96030.95660.99910.96680.93280.93050.96680.96090.96480.9625
DenseNet2010.95210.94620.95240.94050.95390.94030.95040.94250.95000.93500.92530.91910.95020.94440.99670.95840.96680.96090.92310.91700.96150.9543
Xception0.95390.95210.95760.94740.94890.94030.94640.94310.95230.94170.92450.92010.94790.94230.99780.96170.96480.96250.96150.95430.91480.9140
Table 12. The performance of TREE on the combined dataset.
Table 12. The performance of TREE on the combined dataset.
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.97090.80130.98720.89380.98810.88630.98660.86960.98550.89810.98100.83410.98690.87910.99440.94310.99270.94170.99290.91810.98960.9264
ResNet180.98720.89300.97590.81000.98700.86910.98600.86870.98530.87650.98170.82710.98330.84980.99270.94030.99160.92680.99040.89600.99100.9144
ResNet500.98810.88420.98730.87200.97910.79740.98680.86790.98720.87380.98150.80090.98730.86450.99260.93380.99230.92300.99010.89670.99060.8993
ResNet1010.98660.87040.98590.87140.98680.86890.97470.77350.98420.87120.98060.81000.98580.86280.99300.93130.99070.92190.99060.89240.98920.9166
DarkNet190.98560.89650.98550.88060.98700.87000.98440.87340.97150.82550.98090.82040.98490.87670.99220.93350.98940.91790.98840.88950.98820.9117
DarkNet530.98110.83470.98210.82710.98130.80010.98040.80760.98090.82140.96980.69020.97930.81530.99080.90910.98950.90220.98580.86550.98540.8722
MobileNetV20.98680.88010.98330.84790.98720.86430.98590.86200.98470.87520.97960.81430.97510.76290.99320.92680.99110.92800.98830.90090.99080.9117
ShuffleNet0.99440.94230.99260.94010.99280.93520.99300.93090.99210.93290.99060.91260.99320.93010.98940.91660.99540.95500.99330.94030.99500.9490
InceptionV30.99280.94070.99160.92480.99210.92400.99080.91990.98940.91740.98930.90420.99110.92640.99540.95520.98540.88810.99250.92870.99340.9321
DenseNet2010.99290.91680.99040.89670.98990.89630.99030.89930.98820.88610.98630.86710.98840.90030.99330.93930.99230.92760.98260.86040.99180.9264
Xception0.98960.92780.99130.91340.99080.89790.98910.91500.98810.91030.98550.87300.99080.91010.99500.94840.99330.93270.99180.92580.98290.8628
Table 13. The performance of KNN on the combined dataset.
Table 13. The performance of KNN on the combined dataset.
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.87850.85240.93980.92910.93570.92440.93070.91720.93610.92520.89920.87750.93350.91620.97590.97210.96620.96130.95230.93820.95790.9490
ResNet180.93980.92910.88810.86510.92580.91070.92590.90890.93160.91640.89840.88030.91970.90280.96960.96430.95890.94900.94130.92720.95270.9431
ResNet500.93570.92440.92580.91070.87840.85390.92100.90730.92400.91070.89110.86630.92730.91050.97050.96330.95440.94430.94350.93380.94310.9352
ResNet1010.93070.91720.92590.90890.92100.90730.86080.83390.91700.89650.88970.86570.91760.90220.96900.96450.95690.94780.94470.93070.94750.9360
DarkNet190.93610.92520.93160.91640.92400.91070.91700.89650.89900.87690.89980.88160.92430.90810.96190.95170.92860.91520.91960.90380.93890.9248
DarkNet530.89920.87750.89840.88030.89110.86630.88970.86570.89980.88160.80700.75130.88980.86450.95710.94780.94160.93150.92010.90500.92730.9117
MobileNetV20.93350.91620.91970.90280.92730.91050.91760.90220.92430.90810.88980.86450.86100.83040.96930.96470.95780.94880.94630.93820.94940.9390
ShuffleNet0.97590.97210.96960.96430.97050.96330.96900.96450.96190.95170.95710.94780.96930.96470.95630.94680.97870.97150.97230.96560.97640.9709
InceptionV30.96620.96130.95890.94900.95440.94430.95690.94780.92860.91520.94160.93150.95780.94880.97870.97150.93740.92480.96200.95600.96490.9590
DenseNet2010.95230.93820.94130.92720.94350.93380.94470.93070.91960.90380.92010.90500.94630.93820.97230.96560.96200.95600.92000.90160.96030.9503
Xception0.95790.94900.95270.94310.94310.93520.94750.93600.93890.92480.92730.91170.94940.93900.97640.97090.96490.95900.96030.95030.92070.9005
Table 14. Overall comparison of success with different neuron counts on combined datasets.
Table 14. Overall comparison of success with different neuron counts on combined datasets.
MetricsThe Number of Neurons in the Hidden Layer of RWNSVMTREEKNN
5010025050010002000
The Best Training Accuracy0.96430.97060.97680.98160.98810.99690.99910.99540.9787
The Best Test Accuracy0.96300.96850.97250.97440.97440.97230.96680.95500.9721
Average Training Accuracy0.91850.92620.93420.94200.95510.97830.94500.98770.9358
Average Test Accuracy0.91660.92270.92610.92590.92120.90140.93500.88810.9224
Table 15. The performance analysis of RWN on the combined dataset (Activation Function: Sigmoid).
Table 15. The performance analysis of RWN on the combined dataset (Activation Function: Sigmoid).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.87580.85930.94090.93400.93330.92680.93070.92310.93900.93060.90000.89190.93260.92510.97420.97030.96910.96490.95180.94450.95650.9512
ResNet180.94100.93360.88430.87550.92580.91540.92350.91400.93260.92400.89600.88560.91770.90880.96790.96300.95850.95240.94080.93380.95220.9461
ResNet500.93340.92650.92560.91480.87510.86400.92150.91130.93010.91940.88830.87540.92360.91490.96790.96300.95590.94920.94260.93430.94470.9382
ResNet1010.93060.92300.92340.91380.92140.91090.86140.85120.92550.91470.88090.87060.91880.91080.96470.95960.95750.95160.94250.93430.94730.9396
DarkNet190.93910.93090.93260.92320.93030.92020.92530.91520.89360.88210.89470.88350.92880.91940.96430.95790.93980.93120.92680.91740.94320.9343
DarkNet530.90010.89170.89630.88700.88800.87670.88120.87030.89470.88300.78220.76120.88790.87710.95480.94950.94040.93440.91650.90940.92640.9190
MobileNetV20.93260.92590.91780.90920.92350.91500.91870.91090.92900.91940.88780.87760.85250.84060.96850.96420.95900.95360.94430.93840.94930.9424
ShuffleNet0.97410.97030.96800.96350.96780.96320.96470.95940.96440.95800.95490.94870.96860.96440.95390.94830.97740.97400.96900.96430.97550.9712
InceptionV30.96910.96460.95850.95230.95610.95000.95730.95150.93970.93090.94040.93380.95920.95340.97730.97350.93940.93130.96260.95670.96650.9613
DenseNet2010.95190.94470.94090.93370.94220.93530.94260.93520.92690.91820.91680.90950.94440.93820.96900.96450.96260.95670.92160.91190.96050.9550
Xception0.95650.95120.95220.94610.94470.93840.94730.93950.94310.93550.92640.91880.94910.94230.97550.97150.96630.96190.96050.95520.92360.9121
Table 16. The performance analysis of RWN on the combined dataset (Activation Function: Sine).
Table 16. The performance analysis of RWN on the combined dataset (Activation Function: Sine).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.86890.86060.81120.78240.83390.80530.87150.84910.59630.50440.87050.84870.91490.90270.59640.50330.95020.93970.90740.89100.94300.9332
ResNet180.81170.78300.77740.74500.69980.65470.73290.69240.59680.50360.77410.74240.75740.72100.59640.50560.76880.73380.74640.70620.78190.7505
ResNet500.83390.80800.69680.65030.78690.75520.76790.73230.59620.50280.77790.74400.79050.75810.59620.50370.80490.77510.78230.74880.81070.7804
ResNet1010.87240.84990.73350.69480.76720.73050.82660.80180.59600.50490.80210.76900.83010.80290.59600.50270.85470.83040.82300.79480.85730.8327
DarkNet190.59660.50580.59580.50250.59600.50500.59640.50180.59660.50490.59630.50230.59630.50360.59610.50220.59640.50200.59640.50380.59610.5008
DarkNet530.87000.84710.77360.73820.77780.74250.80060.76730.59610.50530.77050.73750.83640.81090.59620.50190.88230.85950.84540.81970.87560.8526
MobileNetV20.91470.90230.75750.72230.78980.75760.83040.80420.59630.50200.83600.81080.84730.83330.59620.50340.90560.88870.86220.83890.90580.8904
ShuffleNet0.59560.50370.59600.50240.59670.50150.59690.50290.59670.50070.59660.50250.59610.50280.59630.50310.59660.50490.59660.50230.59710.5009
InceptionV30.95100.94190.76770.73250.80640.77590.85510.82880.59630.50520.88260.86020.90530.88870.59610.50380.93170.92460.87770.85620.92350.9089
DenseNet2010.90710.89000.74470.70680.78270.74840.82180.79410.59660.50380.84540.82050.86140.83940.59640.50160.87460.85250.88620.87050.88120.8603
Xception0.94330.93440.77890.74460.81020.78160.85760.83420.59620.50400.87630.85330.90510.89030.59630.50180.92300.90890.88160.86080.91740.9082
Table 17. The performance analysis of RWN on the combined dataset (Activation Function: Tribas).
Table 17. The performance analysis of RWN on the combined dataset (Activation Function: Tribas).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.87850.85160.92410.91460.91710.90750.91660.90760.65260.59540.87920.87020.92810.92070.93290.91880.96500.95930.92870.91320.95450.9466
ResNet180.92430.91460.86740.85660.89160.87710.88740.87450.72620.68380.85090.83650.89860.88630.91220.89730.93590.92680.92140.91060.93690.9271
ResNet500.91740.90800.89180.87750.85440.84360.88470.87050.72770.68440.83950.82270.90160.89110.91240.89600.93570.92670.92110.91110.92780.9184
ResNet1010.91640.90600.88760.87480.88530.87200.84610.83640.70060.65250.83680.82080.89640.88680.90310.88650.93630.92770.92050.90990.92940.9198
DarkNet190.65240.59610.72600.68320.72660.68460.70150.65600.63740.57580.71170.66560.70130.65350.73120.69250.68390.63330.64590.58550.69750.6476
DarkNet530.87950.86940.85110.83490.83930.82300.83740.82070.71200.66750.75330.73800.85390.84050.88850.87160.91070.90170.88720.87740.90400.8947
MobileNetV20.92840.92030.89860.88670.90200.88960.89670.88580.70020.65410.85400.84170.84950.84000.92600.91160.95110.94410.93580.92750.94390.9367
ShuffleNet0.93300.91840.91290.89860.91310.89870.90360.88740.73180.68930.88810.87180.92590.91190.89990.88200.93330.92030.92730.91410.93640.9235
InceptionV30.96490.95880.93610.92740.93550.92750.93590.92740.68430.63550.91030.90130.95150.94510.93350.91930.93220.92430.95400.94540.96210.9557
DenseNet2010.92890.91330.92130.91030.92070.90980.92100.91020.64570.58510.88770.87780.93580.92740.92690.91210.95380.94530.86800.84410.95360.9454
Xception0.95440.94630.93700.92720.92800.91980.92940.92000.69750.64920.90380.89410.94370.93690.93560.92280.96200.95530.95370.94600.91850.9068
Table 18. The performance analysis of RWN on the combined dataset (Activation Function: Radbas).
Table 18. The performance analysis of RWN on the combined dataset (Activation Function: Radbas).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.87690.85580.93250.92360.92460.91520.92430.91550.67800.62540.88850.87900.93080.92160.95480.94490.96750.96230.94120.92880.95620.9494
ResNet180.93240.92390.87490.86540.90630.89290.90330.89130.76680.73010.86800.85400.90810.89760.94080.92870.94740.93900.93080.92120.94520.9369
ResNet500.92480.91550.90590.89380.86320.85060.90090.88810.76610.72980.85720.84060.91170.90070.94010.92850.94470.93650.92990.91970.93570.9261
ResNet1010.92410.91480.90310.89060.90060.88700.85400.84260.73650.69600.85520.84000.90740.89670.93130.91920.94600.93770.93030.92100.93800.9276
DarkNet190.67740.62490.76550.72950.76650.73200.73630.69590.65180.59360.74890.71000.73610.69570.77300.73680.71810.67460.66690.61400.73280.6931
DarkNet530.88890.87920.86830.85420.85710.84040.85530.83970.74810.71110.76360.74650.86660.85430.91580.90300.92320.91390.89900.88930.91440.9055
MobileNetV20.93090.92240.90790.89790.91130.90020.90720.89580.73550.69540.86630.85360.85210.83990.94910.93890.95510.94860.93950.93260.94690.9388
ShuffleNet0.95430.94540.94080.92800.93970.92920.93180.91920.77350.73990.91540.90230.94900.93840.92840.91500.95600.94630.94950.93970.95840.9494
InceptionV30.96760.96180.94720.93870.94470.93570.94610.93750.71700.67200.92320.91490.95520.94870.95620.94620.93570.92660.95800.95100.96490.9582
DenseNet2010.94150.92860.93080.92140.93060.92000.93010.92020.66740.61400.89920.89000.93980.93200.94930.93880.95780.95150.89210.87240.95680.9490
Xception0.95620.94930.94560.93660.93570.92710.93810.92880.73280.69230.91440.90520.94680.93850.95840.94940.96470.95870.95680.94950.92320.9079
Table 19. The performance analysis of RWN on the combined dataset (Activation Function: Hardlim).
Table 19. The performance analysis of RWN on the combined dataset (Activation Function: Hardlim).
Feature Extracted ModelsEfficientNetResNet18ResNet50ResNet101DarkNet19DarkNet53MobileNetV2ShuffleNetInceptionV3DenseNet201Xception
TrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTestTrainTest
EfficientNet0.88190.85210.93690.93010.92980.92420.92800.92030.93740.92810.89360.88740.93080.92390.97300.96850.96780.96290.95060.94190.95660.9518
ResNet180.93680.93020.87880.87240.92020.91100.91810.90980.93130.92180.88800.87990.91330.90570.96660.96200.95400.94790.93660.92960.94930.9435
ResNet500.93000.92350.92000.91130.86890.86140.91530.90720.92840.91830.87850.86930.91830.91150.96640.96170.95170.94630.93760.93050.94150.9363
ResNet1010.92810.92100.91820.90930.91610.90720.85520.84750.92330.91180.87320.86330.91320.90480.96330.95770.95290.94710.93770.93160.94370.9370
DarkNet190.93750.92820.93110.92140.92830.91720.92300.91250.89360.88260.89400.88160.92750.91670.96340.95650.93720.92780.92490.91520.94150.9328
DarkNet530.89320.88660.88800.87950.87880.86840.87320.86350.89400.88150.76230.75080.87720.86890.95270.94780.93240.92750.90840.90190.92120.9160
MobileNetV20.93100.92360.91320.90620.91850.91180.91330.90570.92720.91650.87700.86880.85060.84240.96730.96310.95550.95060.94100.93520.94690.9414
ShuffleNet0.97300.96870.96650.96150.96640.96150.96320.95770.96320.95690.95260.94670.96730.96270.95340.94730.97590.97120.96790.96310.97440.9704
InceptionV30.96780.96270.95390.94860.95150.94670.95280.94780.93740.92780.93240.92770.95560.95030.97600.97210.93310.92600.96030.95480.96500.9600
DenseNet2010.95070.94200.93690.92990.93750.93090.93770.93090.92510.91470.90860.90220.94120.93550.96790.96290.96040.95470.91810.90570.95850.9524
Xception0.95660.95060.94930.94390.94170.93670.94350.93690.94150.93270.92110.91560.94680.94110.97460.97000.96490.96020.95850.95240.91870.9095
Table 20. Comparisons of activation functions used in RWN on the combined dataset.
Table 20. Comparisons of activation functions used in RWN on the combined dataset.
SigTansigSinTribasRadbasHardlimit
The Best Training Accuracy0.97740.97680.95100.96500.96760.9760
The Best Test Accuracy0.97400.97250.94190.95930.96230.9721
The Average Training Accuracy0.93540.93420.75680.87320.88890.9319
The Average Test Accuracy0.92770.92610.70880.85520.87300.9246
Table 21. Performance analysis of the proposed method compared to state-of-the-art classifiers.
Table 21. Performance analysis of the proposed method compared to state-of-the-art classifiers.
MetricsRWNCatBoostDecision TreeGaussian Naïve BayesGradient BoostingKNNLightGBMLogistic RegressionRandom ForestRidge ClassifierSVMXGBoost
The Best Training Accuracy0.98490.95490.99690.94370.94720.98021.00000.97340.99670.96340.99961.0000
The Best Test Accuracy0.98430.9510.96460.96460.94510.98230.96070.95480.95480.96080.97840.9568
Average Training Accuracy0.98160.95350.99540.94270.94360.97871.00000.97190.99610.96190.99911.0000
Average Test Accuracy0.97430.94460.95520.94250.9340.97150.94840.94600.94250.94840.96680.9486
Mean Test F1-Score0.97770.94510.95480.94420.93480.97170.94890.94640.94240.94910.96670.9491
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kiran, M.S.; Seyfi, G.; Yilmaz, M.; Esme, E.; Wang, X. Feature Fusion Using Deep Learning Algorithms in Image Classification for Security Purposes by Random Weight Network. Appl. Sci. 2025, 15, 9053. https://doi.org/10.3390/app15169053

AMA Style

Kiran MS, Seyfi G, Yilmaz M, Esme E, Wang X. Feature Fusion Using Deep Learning Algorithms in Image Classification for Security Purposes by Random Weight Network. Applied Sciences. 2025; 15(16):9053. https://doi.org/10.3390/app15169053

Chicago/Turabian Style

Kiran, Mustafa Servet, Gokhan Seyfi, Merve Yilmaz, Engin Esme, and Xizhao Wang. 2025. "Feature Fusion Using Deep Learning Algorithms in Image Classification for Security Purposes by Random Weight Network" Applied Sciences 15, no. 16: 9053. https://doi.org/10.3390/app15169053

APA Style

Kiran, M. S., Seyfi, G., Yilmaz, M., Esme, E., & Wang, X. (2025). Feature Fusion Using Deep Learning Algorithms in Image Classification for Security Purposes by Random Weight Network. Applied Sciences, 15(16), 9053. https://doi.org/10.3390/app15169053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop