Next Article in Journal
EvTransPose: Towards Robust Human Pose Estimation via Event Camera
Next Article in Special Issue
An IoT Featureless Vulnerability Detection and Mitigation Platform
Previous Article in Journal
High-Voltage Electrostatic Discharge/Electrical Overstress Co-Protection Implementing Gradual-Triggered SCR and MOS-Stacked Configuration
Previous Article in Special Issue
Video-Wise Just-Noticeable Distortion Prediction Model for Video Compression with a Spatial–Temporal Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-MCKD: Enhancing the Effectiveness and Efficiency of Knowledge Transfer in Malware Classification

Division of Software, Yonsei University, Wonju-si 26493, Gangwon-do, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(6), 1077; https://doi.org/10.3390/electronics14061077
Submission received: 30 January 2025 / Revised: 27 February 2025 / Accepted: 6 March 2025 / Published: 8 March 2025

Abstract

:
As malware continues to evolve, AI-based malware classification methods have shown significant promise in improving the malware classification performance. However, these methods lead to a substantial increase in computational complexity and the number of parameters, increasing the computational cost during the training process. Moreover, the maintenance cost of these methods also increases, as frequent retraining and transfer learning are required to keep pace with evolving malware variants. In this paper, we propose an efficient knowledge distillation technique for AI-based malware classification methods called Self-MCKD. Self-MCKD transfers output logits that are separated into the target class and non-target classes. With the separation of the output logits, Self-MCKD enables efficient knowledge transfer by assigning weighted importance to the target class and non-target classes. Also, Self-MCKD utilizes small and shallow AI-based malware classification methods as both the teacher and student models to overcome the need to use large and deep methods as the teacher model. From the experimental results using various malware datasets, we show that Self-MCKD outperforms the traditional knowledge distillation techniques in terms of the effectiveness and efficiency of its malware classification.

Graphical Abstract

1. Introduction

With the rapid development of the Internet, computers have become an integral part of people’s daily lives. However, this growth has also led to a gradual increase in the number of malware, which has become one of the major threats to computer security. Some hackers exploit vulnerabilities in computers to steal data or obtain private information without the user’s permission. As a representative example, the WannaCry malware caused billions of dollars in damages to millions of computer devices [1].
To address such threats in a timely and effective manner, many countermeasures have been proposed. Traditionally, many studies have focused on malware classification methods using a signature-based approach, relying on statistical methods. However, as malware continues to evolve with new variants, these methods have become less effective, leading to a decrease in the classification performance.
In response, some studies have shifted towards AI-based methods, which have shown significant promise in improving the malware classification performance [2]. Until recently, AI-based malware classification methods focused on enhancing their performance by increasing the model size and depth. However, these methods lead to a substantial rise in computational complexity and the number of parameters, significantly raising the computational cost during the training process. Moreover, the maintenance cost of AI-based malware classification methods also increases, as frequent retraining and transfer learning are required to keep up with evolving malware variants. These challenges not only increase the training time of AI-based malware classification methods but also increase the inference time. For example, such challenges present practical limitations when applying AI-based malware classification methods to IoT networks in industrial CPSs, which are resource-constrained environments [3]. 
To address these challenges, there is a growing need for research on lightweight techniques for AI-based malware classification methods. The existing lightweight techniques for AI-based malware classification methods are grouped into two categories of architectures, i.e., model compaction techniques and model optimization techniques. Model compaction techniques simplify the model structure by replacing layers with more efficient ones for lightweight AI-based malware classification methods. However, these methods suffer from computational complexity and long inference times. On the other hand, model optimization techniques reduce the computational complexity and inference time by optimizing the computational complexity of AI-based malware classification methods. In particular, knowledge distillation, which is a representative model optimization technique, has been actively studied as an efficient lightweight technique for AI-based malware classification methods.
Knowledge distillation allows small and shallow AI-based malware classification methods, as the student model, to mimic large and deep AI-based malware classification methods, as the teacher model, through knowledge transfer. Unfortunately, previous studies on knowledge distillation techniques for AI-based malware classification methods have still faced issues in terms of the computational complexity and inefficient knowledge transfer. Also, they require large and deep AI-based malware classification methods for the teacher model, which leads to high computational costs during the training process.
To solve these challenges, we propose an efficient knowledge distillation technique for AI-based malware classification methods called Self-MCKD. As shown in Figure 1, different from the traditional knowledge distillation techniques for AI-based malware classification methods, which transfer combined output logits composed of the target class and non-target classes, Self-MCKD transfers output logits that are separated out into these parts. With the separation of the output logits, Self-MCKD enables efficient knowledge transfer by assigning weighted importance to the target class and non-target classes. Also, Self-MCKD only uses small and shallow methods as both the teacher and student models to overcome the need to use large and deep methods as the teacher model in knowledge distillation techniques for AI-based malware classification methods. Since Self-MCKD enables efficient knowledge transfer, it not only enhances the performance of small and shallow AI-based malware classification methods but also enables them to surpass large and deep methods. Also, it accelerates the convergence speed and reduces the training time for small and shallow AI-based malware classification methods.
The main contributions of this paper can be summarized as follows:
  • We propose Self-MCKD, a novel knowledge distillation technique for AI-based malware classification methods. Self-MCKD separates the output logits into the target class and non-target classes, assigning a weighted importance factor that maximizes the effectiveness of knowledge transfer.
  • To enhance the knowledge transfer efficiency, we introduce small and shallow AI-based malware classification methods as both the teacher and student models in Self-MCKD.
  • From the experimental results using representative benchmark datasets, we show that Self-MCKD significantly improves the performance of a small and shallow baseline model, even surpassing that of a large and deep baseline model. Also, we show that Self-MCKD outperforms previous knowledge distillation techniques in both its effectiveness and efficiency. Specifically, with the same FLOPs, MACs, and parameters, Self-MCKD achieves improvements in accuracy, recall, precision, and F1-score of 0.2% to 1.4% while reducing the training time by 20 s.

2. Related Works

In this section, we first overview previous AI-based malware classification methods and lightweight techniques for AI-based malware classification methods, and then we introduce the most effective knowledge-distillation-based technique for AI-based malware classification methods.

2.1. AI-Based Malware Classification Methods

In the field of malware classification, many previous methods have utilized executable files represented by opcodes or assembly codes [4]. However, these methods faced several limitations, such as difficulty analyzing obfuscated code, the time-consuming nature of static analyses, and the high resource demands of dynamic analyses [5]. To address these limitations, several studies have applied AI-based methods, specifically machine learning (ML) and deep learning (DL), after converting executable files into gray-scale images [6]. Nataraj et al. were the first to apply the K-Nearest Neighbors (KNN) algorithm, an ML technique, to executable files converted into gray-scale images for malware classification [7]. To achieve an improved performance over a KNN algorithm, Hardy et al. proposed an SVM-based malware classification method [8]. Also, Drew et al. leveraged modern gene classification tools [9], and Ahmadi et al. leveraged XGBoost techniques [10], to progressively improve the performance of malware classification with more advanced ML models. However, as the variety of malware variants increased, the complexity of malware analysis also increased. Moreover, this revealed certain limitations of ML-based malware classification methods, especially their inability to automatically extract and learn nuanced features of malware variants.
To address the limitations of ML-based malware classification methods, DL-based malware classification methods, which can automatically extract nuanced features, have emerged. In particular, convolutional neural network (CNN)-based malware classification methods have gained prominence due to their superior image analysis capabilities when applied to executable files converted into gray-scale images. Daniel et al. [11] applied a shallow, simple CNN model for malware classification using gray-scale images of malware. They also showed that DL-based malware classification methods outperformed previous ML-based malware classification methods. Following this study, Mahmoud et al. hypothesized that deeper CNN architectures could further improve the performance in malware classification [12]. Based on this assumption, they enhanced the performance of DL-based malware classification methods using the VGG16 model. To achieve an even better classification performance, Pratikkumar et al. proposed a migration learning framework for malware classification, utilizing more advanced models such as the ResNet52 and VGG19 models [13].

2.2. Lightweight Techniques for DL-Based Malware Classification Methods

As DL-based malware classification methods have become increasingly complex and deeper, the associated hardware burden has also increased. Also, since this has led to longer training and inference times, several recent studies have focused on developing lightweight DL-based malware classification methods. In this section, we summarize and discuss key studies related to the lightweight techniques for DL-based malware classification methods. Early lightweight techniques for DL-based malware classification methods focused on simplifying the model structure, also known as model compaction techniques. Specifically, Fathurrahman et al. [14] proposed a lightweight malware classification method by replacing the deep structures of the CNN model with residual layers and global average pooling. This approach outperformed the lightweight CNN-based malware classification method proposed by Su et al. [15], offering both reduced complexity and an improved performance. However, model compaction techniques, which simply replace layer structures, still suffer from computational complexity and long inference times [16,17,18].
To address these challenges, efficient lightweight techniques, also known as model optimization techniques, have emerged. These techniques, such as weight pruning [19], model quantization [20], and knowledge distillation [21], reduced the computational complexity and inference time in DL-based malware classification methods. In particular, knowledge distillation has been actively studied as an efficient lightweight technique for DL-based malware classification methods. Zhi et al. proposed a knowledge-distillation-based technique for recurrent neural network (RNN)-based malware classification methods [22]. Specifically, they enhanced the performance of an RNN-based malware classification method (the student model) using a transformer-based malware classification method as the teacher model. Xia et al. [23] also proposed a lightweight technique for multi-layer perceptron (MLP)-based malware classification methods using knowledge distillation. Different from Zhi et al.’s technique, which uses different families of models for the teacher and student models, Xia et al.’s technique uses the same family of MLP models for both. By using the same MLP family of models, they improved the performance of the student MLP-based malware classification method to a level comparable to that of the teacher MLP-based model. Building on this foundation, recent studies have further combined effective learning methods with knowledge distillation, such as the two-dimensional Fourier transform and federated learning [24,25].
Previous knowledge-distillation-based techniques for DL-based malware classification methods showed that small and shallow student models could achieve a performance comparable to that of large and deep teacher models. However, despite the promising performance of these techniques, they still suffered from computational complexity and inefficient knowledge transfer. Moreover, many of these techniques relied on using large and deep DL-based methods as the teacher model, which incurred a high computational cost during the training process. These problems have motivated us to seek a new knowledge-distillation-based technique for DL-based malware classification methods which enables efficient knowledge transfer with minimal computational cost and complexity.

3. Preliminary

Knowledge Distillation

Knowledge distillation was first introduced by Bucila et al. for model compression [26] and gained popularity after Hinton et al. generalized it [21]. Knowledge distillation is a model optimization technique that allows the small and shallow student model to mimic the performance of a large and deep teacher model. Specifically, the student model S is trained to mimic the performance of the teacher model T by leveraging the teacher’s “Dark Knowledge”. Here, “Dark Knowledge” refers to the softened class probabilities predicted by the teacher model, which are distilled into the student model. By utilizing this “Dark Knowledge”, the student model improves its performance by learning not only from the original training data but also by incorporating the nuanced knowledge provided by the teacher’s predictions. The process of distilling this “Dark Knowledge”, which is called the knowledge distillation process ( K D ), can be expressed as follows:
K D = x i X KL s o f t m a x f T ( x i ) τ , s o f t m a x f S ( x i ) τ ,
where X represents the training dataset, x i represents individual data points, and f T ( x i ) and f S ( x i ) represent the class probabilities predicted by the teacher and student models, respectively. Additionally, τ is the temperature parameter used to soften the class probabilities. The Kullback–Leibler (KL) divergence is used as the loss function to measure the difference between the softened outputs of the teacher and student models. By minimizing the KL divergence, the student model is trained to mimic the behavior of the teacher model, effectively absorbing the teacher’s “Dark Knowledge” and enhancing its own classification performance. Knowledge distillation can be implemented in two main ways: using logit-based and feature-based approaches. In Figure 2, we show the differences between these two knowledge distillation approaches.
As shown in Figure 2a, the feature-based approach, which was first proposed by Romero et al. [27], distills knowledge to the student model from the teacher model’s hint layers. This approach focuses on training the student to mimic the output of the intermediate layers. Specifically, the feature-based approach transfers the rich knowledge of the teacher to the student, enabling the student to achieve a performance similar to that of the teacher [28,29]. However, since the knowledge distilled from the hint layers has a high computational complexity, the feature-based approach requires a significant training time for the student model. Also, the feature-based approach is inefficient for the lightweight techniques of DL-based malware classification methods because DL-based malware classification methods require frequent training and transfer learning.
On the other hand, the logit-based approach distills knowledge to the student model from the teacher model’s output layer, which produces softened class probabilities, as shown in Figure 2b. Different from the feature-based approach, this approach focuses on training the student to mimic the output logits of the teacher. Also, since the logit-based approach has low computational complexity, it may be a solution to the long training time problem in the feature-based approach. For example, traditional knowledge distillation, which is a representative logit-based approach, enables a reduced training time for the student model compared to that in the feature-based approach. Unfortunately, the existing logit-based approaches suffer from inefficient knowledge transfer, which limits the performance of the student model. Therefore, we propose a new logit-based approach which can transfer knowledge more efficiently to the student model compared to the existing logit-based approaches.

4. The Proposed Method

In this section, we overview the operation of Self-MCKD in detail. First, we introduce the difference between traditional knowledge distillation and Self-MCKD, and then we reformulate traditional knowledge distillation for Self-MCKD. Next, we describe the network architecture in Self-MCKD used for both the teacher and student models.

4.1. Revisiting Traditional Knowledge Distillation

Traditional knowledge distillation transfers the combined knowledge on the target and non-target classes to the student model. However, this approach suffers from inefficient knowledge transfer, as the combined knowledge is often emphasized in an unbalanced manner, reducing its effectiveness [30]. Also, traditional knowledge distillation tends to use high-performing large and deep teacher models, which require a high computational cost during the training process. In the field of malware classification, effective knowledge transfer is essential for better generalization against evolving malware variants. However, traditional knowledge distillation has limitations in this domain due to its unbalanced knowledge transfer. Moreover, its reliance on large and deep teacher models incurs high computational costs, making frequent retraining challenging and limiting its adaptability to newly emerging malware threats.
To address these problems in the field of malware classification, we propose a novel knowledge distillation technique called Self-MCKD. Different from traditional knowledge distillation, Self-MCKD transfers the knowledge on the target and non-target classes independently. Also, Self-MCKD does not rely on a large and deep teacher model. Instead, it utilizes a pre-trained small and shallow student model to serve as the teacher, transferring knowledge to the untrained student model during the training process [31].

4.2. Notations for Self-MCKD

In this section, we define two types of output probabilities—target malware class probabilities and non-target malware class probabilities—to reformulate the knowledge transfer process in traditional knowledge distillation. When a malware sample of the i-th class is given, the output probabilities of the malware classification model are defined as p = [ p 1 , p 2 , p 3 , , p i , , p M ] R 1 × M , where p i represents the output probabilities of the i-th malware class, and M is the number of malware families. Here, p i is calculated as follows:
p i = exp ( z i ) j = 1 M exp ( z j ) ,
where z i is the output logit for the i-th malware class in the malware classification model.
Next, to separate the output probabilities related to the target malware class and non-target malware classes, we define binary probabilities as bm = [ p t m , p n t m ] R 1 × 2 . Here, p t m and p n t m can be expressed into
p t m = exp ( z t m ) j = 1 M exp ( z j ) , p n t m = k = 1 , k t m M exp ( z k ) j = 1 M exp ( z j ) ,
where p t m represents the probabilities of the target malware class, and p n t m represents the probabilities of all of the other non-target malware classes. Meanwhile, we define p ^ = [ p ^ 1 , , p ^ i 1 , p ^ i + 1 , , p ^ M 1 ] R 1 × ( M 1 ) , which means the output probabilities of the non-target malware classes, i.e., excluding the i-th malware class. Here, p ^ i is calculated as follows:
p ^ i = exp ( z i ) j = 1 , j t m M exp ( z j ) .

4.3. Reformulating Knowledge Distillation for Self-MCKD

Based on the above notations, we reformulate the knowledge distillation process ( K D ) from Equation (1) as follows:
K D = K L ( p T | | p S ) = p t m T log ( p t m T p t m S ) + i = 1 , i t m M p i T log ( p i T p i S ) ,
where p T represents the output probabilities of the teacher model, and p S represents the output probabilities of the student model.
Different from the loss function in traditional knowledge distillation, which aims to simply reduce the overall difference in the probabilities between the teacher and student models, Self-MCKD aims to reduce the difference for both the target and non-target malware classes separately between the teacher and student models. Also, since Self-MCKD replaces the large and deep teacher model with a pre-trained small and shallow student model, p T can be replaced with p t S . Finally, considering that p ^ equals p i / p n t m , we can reformulate Equation (5) as follows:
K D = p t m t S log p t m t S p t m S + i = 1 , i t m M p n t m t S p ^ i t S log p ^ i t S p ^ i S + log p n t m t S p n t m S = p t m t S log p t m t S p t m S + p n t m t S log p n t m t S p n t m S + p n t m t S i = 1 , i t m M p ^ i t S log p ^ i t S p ^ i S = K L ( bm t S | | bm S ) + p n t m t S K L ( p ^ t S | | p ^ S ) .
After reformulating Equation (6), it can be written as follows:
K D = K L ( bm t S | | bm S ) + p n t m t S K L ( p ^ t S | | p ^ S ) .
As shown in Equation (7), the loss function from traditional knowledge distillation can be reformulated as a weighted sum of two terms. Here, K L ( bm t S | | bm S ) represents the knowledge of the target malware class between the teacher and student models, which we refer to as (Target Malware Class Knowledge Distillation) TMKD. On the other hand, K L ( p ^ t S | | p ^ S ) represents the knowledge of the non-target malware classes between the teacher and student models, which we refer to as (Non-Target Malware Class Knowledge Distillation) NMKD. These separated types of knowledge in Self-MCKD are shown in Figure 3. According to Zhao et al.’s work [30], the p n t m t S factor, derived from traditional knowledge distillation, enforces an emphasis on the non-target malware classes. This enforced emphasis prevents autonomous adjustment of the importance of the target and non-target malware classes. As a result, it restricts the effectiveness of the knowledge transfer to student models. To address the restrictive effect of the p n t m t S factor, we introduce a factor β . Also, to place proper emphasis on TMKD, we introduce a factor α . Consequently, these factors enable Self-MCKD to emphasize TMKD and NMKD, respectively. By applying these factors and making the necessary substitutions into Equation (7), the loss function in Self-MCKD can be defined as shown in Equation (8) and calculated using Algorithm 1.
Self - MCKD = α TMKD + β NMKD .
Algorithm 1 Self-MCKD loss function pseudo-code
Input: Student logits z S , pre-trained student logits z t S , target malware y , factors α , β , temperature τ
Output: Self-MCKD Loss
1. Compute the target malware and non-target malware masks:
GT _ Mask One - hot ( y )
Other _ Mask 1 GT _ Mask
2. Normalize logits with temperature:
p S Softmax ( z S / τ )
p t S Softmax ( z t S / τ )
3. Concatenate the probabilities:
p s Concat ( p s · GT _ Mask , p s · Other _ Mask )
p t S Concat ( p t S · GT _ Mask , p t S · Other _ Mask )
4. Compute the TMKD loss:
TMKD KL ( log ( p s ) , p t S ) · τ 2
5. Compute the NMKD loss:
NMKD KL ( log ( p s GT _ Mask )
p t S GT _ Mask ) · τ 2
6. Return the total loss:
Self - MCKD α TMKD + β NMKD

5. Evaluation Results

To demonstrate the effectiveness and efficiency of the proposed Self-MCKD in the field of malware classification, we evaluated its performance across various datasets. Specifically, we aimed to evaluate the lightweighting effect of Self-MCKD when applied to representative DL-based malware classification methods in different dataset environments. To this end, we tried to answer the following questions:
  • How do the factors α of TMKD and β of NMKD influence the performance of Self-MCKD?
  • Does Self-MCKD show a good performance on small and shallow malware classification models using various malware datasets?
  • Does Self-MCKD show a better performance than that of the other knowledge distillation methods?

5.1. Experimental Environments

In this section, we describe the experimental environments, including the dataset, model configuration, evaluation metric, and so on.
Dataset: We conducted the experiments using the Malimg dataset (Malimg) [32] and the Microsoft Malware Classification Challenge Dataset (MMCC) [33], both of which are representative datasets for malware classification. The Malimg, originally compiled by Nataraj et al. [7], consists of 9339 malware samples represented as gray-scale images, corresponding to 25 malware families. These families include “Adialer”, “Allaple”, “Lolyda”, “C2LOP”, “Swizzor”, and several others. The MMCC dataset, provided by Microsoft Malware Classification Challenge, consists of 21,741 malware samples categorized into 9 malware families, namely “Ramni”, “Lollipop”, “Kelihos_ver3”, “Vundo”, “Simda”, “Tracur”, “Kelihos_ver1”, “Obfuscator.ACY”, and “Gatak”. As with Malimg, the distribution of the malware samples across classes in the MMCC dataset is not uniform, with some families having significantly more samples than others. In our experiments, we randomly selected 90% of the malware samples from each family for training and used the remaining 10% for testing.
Model configuration: To demonstrate the effectiveness and efficiency of the proposed Self-MCKD, we selected representative DL-based malware classification methods, such as ResNet and VGGNet. In previous malware classification studies [12,34], ResNet and VGGNet achieved a promising classification performance. Thus, we used the ResNet110 and VGG16 models as large and deep baseline models for malware classification. Also, to apply small and shallow baseline models with the proposed Self-MCKD, we selected much lighter homogeneous models, such as the ResNet32 and VGG8 models. For the implementation of these models, we followed the original ResNet and VGGNet architecture [34,35] without modifications. Also, we set these models’ output layers to 9 or 25, depending on the number of malware families in the dataset, and used the stochastic gradient descent (SGD) optimization function to train them. Lastly, we set hyperparameters such as the batch size, epoch, learning rate, momentum, and weight decay to 8, 25, 0.001, 0.9, and 0.0005, respectively.
Evaluation metric: To evaluate the performance of Self-MCKD, we measured the accuracy, recall, precision, and F1-score using different malware datasets. These metrics are derived from a confusion matrix. In the confusion matrix, True Positive (TP) represents correctly predicted samples from the target malware class, while True Negative (TN) represents correctly predicted samples from the non-target malware classes. False Positive (FP) represents non-target malware class samples that are incorrectly predicted as belonging to the target malware class. On the other hand, False Negative (FN) represents target malware class samples incorrectly predicted as belonging to the non-target malware classes. Based on the confusion matrix, accuracy, recall, precision, and F1-score are calculated as follows:
A c c u r a c y = T P + T N F P + T P + F N + T N .
R e c a l l = T P T P + F N .
P r e c i s i o n = T P T P + F P .
F 1 - s c o r e = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l .
Also, to consider the computational efficiency and complexity of tne baseline malware classification models, we measured floating point operations (FLOPs), multiply–accumulate operations (MACs), and the number of model parameters. Here, FLOPs refer to the total number of floating-point operations required during a forward pass, and MACs represent the number of multiply–accumulate operations. Such metrics reflect the computational cost of the convolutional and fully connected layers. In knowledge distillation, both the student and teacher models contribute to the FLOPs and MACs during the inference process. However, the number of parameters, which represents the total trainable parameters of the baseline malware classification models, is measured only for the student model during the training process. We conducted the experiments five times and reported the average results to ensure a reliable evaluation based on these metrics.
The implementation environment: We implemented the baseline malware classification models and knowledge distillation methods using PyTorch version 2.2.2 and Python version 3.11.9. For efficient experiments, the performance was measured on the Windows OS with PyTorch-CUDA version 12.1, an Intel Core i9-12900KF processor, an NVIDIA RTX 3060 GPU, and 64 GB of memory.

5.2. Experimental Analysis

5.2.1. How Do the Factors α of TMKD and β of NMKD Influence the Performance of Self-MCKD?

To evaluate the impact of the factors α and β on the performance of Self-MCKD, we measured the accuracy of the malware classification model while adjusting these factors. Simultaneously, we aimed to verify that these factors provide a better performance compared to that of the p n t m t S factor derived from the traditional knowledge distillation reformulation. To determine the appropriate values for α and β , we conducted experiments by setting their sum to 1 to ensure stability during the training process and prevent degradation of the effectiveness of knowledge transfer. Also, we varied the α and β values from (0.1, 0.9) to (0.9, 0.1).
In Table 1, we show the experimental results of the VGG8 model applied to Self-MCKD using the MMCC and Malimg datasets. According to Table 1, it is observed that too much emphasis on NMKD by increasing the factor β reduced the effectiveness of knowledge transfer. For example, in the experimental results on both datasets, while the VGG8 model with traditional knowledge distillation achieved an accuracy of 98.44% and 99.36%, the VGG8 model with Self-MCKD achieved an accuracy of 98.07% and 99.25% when we set α and β to 0.2 and 0.8, respectively. Also, we observed that too much emphasis on TMKD by increasing the factor α also negatively affected knowledge transfer. For example, in the experimental results on the MMCC dataset, the VGG8 model with Self-MCKD, where α and β were set to 0.9 and 0.1, achieved an accuracy of 98.25%, which was lower than the accuracy of 98.44% achieved by the VGG8 model with traditional knowledge distillation.
On the other hand, when balancing between TMKD and NMKD, we observed that Self-MCKD transferred knowledge more effectively than traditional knowledge distillation. Specifically, when we set both α and β to 0.5, the VGG8 model with Self-MCKD achieved an accuracy of 98.52% and 99.46%, respectively, which was higher than that of the VGG8 model with traditional knowledge distillation. Also, it was observed that a proper balance between TMKD and NMKD maximized the effectiveness of knowledge transfer. For example, the VGG8 model with Self-MCKD, with α and β set to 0.7 and 0.3, achieved an accuracy of 98.62% and 99.57%, respectively, which outperformed the VGG8 model with traditional knowledge distillation.

5.2.2. Does Self-MCKD Show a Good Performance on Small and Shallow Malware Classification Models Using Various Malware Datasets?

To show the effectiveness and efficiency of Self-MCKD on small and shallow malware classification models under various datasets, we measured the computational complexity and accuracy of the ResNet32 and VGG8 models, both with and without Self-MCKD using the MMCC and Malimg datasets. Simultaneously, we aimed to show that the ResNet32 and VGG8 models with Self-MCKD outperformed the larger and deeper malware classification models by comparing their performance to that of the ResNet110 and VGG16 baseline models. Specifically, we measured FLOPs, MACs, and the number of model parameters to assess the computational complexity, while evaluating the performance in terms of accuracy and training time. In Table 2, we show the experimental results for ResNet and VGGNet under the MMCC and Malimg datasets.
From the experimental results, we first observed that the large and deep malware classification models outperformed the small and shallow malware classification models under various datasets. For example, on the MMCC dataset, the ResNet32 and VGG8 baseline models achieved accuracies of 96.87% and 97.70%, respectively, which were lower than the accuracies of 97.79% and 98.25% achieved by the ResNet110 and VGG16 baseline models. Also, on the Malimg dataset, the ResNet110 and VGG16 baseline models achieved accuracies of 99.36% and 99.25%, respectively, while the ResNet32 and VGG8 baseline models both achieved an accuracy of 99.04%. However, deeper models such as the ResNet110 and VGG16 baseline models cost approximately four times more in terms of the FLOPs, MACs, and parameters compared to those of shallower models such as ResNet32 and VGG8. Also, these models required more than twice the training time.
Second, it was observed that the small and shallow malware classification models with Self-MCKD achieved a better performance than that of the large and deep malware classification models, while using only minimal additional computational costs. For example, on the MMCC dataset, while the VGG16 baseline model achieved an accuracy of 98.25% using 627.04 M FLOPs, 313.52 M MACs, and 14.77 M parameters, the VGG8 model with Self-MCKD achieved an accuracy of 98.62% with only 272.02 M FLOPs, 136.01 M MACs, and 3.96 M parameters. Also, while the VGG16 baseline model took 91 min and 8 s to train, the VGG8 model with Self-MCKD required only 28 min and 19 s, which was more than three times faster. Similarly, the ResNet110 baseline model required 508.42 M FLOPs, 254.21 M MACs, and 1.74 M parameters to achieve an accuracy of 97.79%. Its performance was higher than that of the ResNet32 baseline model, which achieved an accuracy of 96.87% with 138.89 M FLOPs, 69.45 M MACs, and 0.47 M parameters. However, the ResNet32 model with Self-MCKD outperformed the ResNet110 baseline model, with 277.78 M FLOPs, 138.90 M MACs, and 0.47 M parameters. For the Malimg dataset, the ResNet32 and VGG8 models with Self-MCKD achieved a comparable performance to that of the ResNet110 and VGG16 baseline models. For example, while the ResNet110 and VGG16 baseline models achieved accuracies of 99.36% and 99.25%, respectively, the ResNet32 and VGG8 models with Self-MCKD achieved accuracies of 99.25% and 99.57%, respectively.
Third, we observed that the small and shallow malware classification models with Self-MCKD effectively transferred knowledge from the teacher to the student models. In other words, Self-MCKD not only improved the classification accuracy of the student model but also enhanced its convergence speed. Specifically, the accuracy of the ResNet32 and VGG8 baseline models improved from 96.87% and 97.70% to 98.53% and 98.62%, respectively, when using Self-MCKD on the MMCC dataset. Similarly, both the ResNet32 and VGG8 baseline models achieved an accuracy of 99.04% without Self-MCKD. On the other hand, for both models with Self-MCKD, the average accuracy increased by approximately 0.3%. Such results show that the ResNet32 and VGG8 models with Self-MCKD maintained the number of parameters while reducing the training time by approximately 1.5 times compared to that for the ResNet32 and VGG8 baseline models.

5.2.3. Does Self-MCKD Show a Better Performance than That of the Other Knowledge Distillation Methods?

To compare the classification performance of Self-MCKD with that of other logit-based knowledge transfer methods, we conducted experiments using Vanilla KD and MLKD. We measured the accuracy, precision, recall, F1-score, and training time of Vanilla KD, MLKD, and Self-MCKD. Here, Vanilla KD refers to a traditional knowledge distillation method that transfers knowledge, combining both the target and non-target classes, from the teacher model to the student model. Also, MLKD refers to Multi-Level Logit Knowledge Distillation (MLKD) [36], a state-of-the-art method that distills knowledge at multiple logit levels. Unlike these approaches, Self-MCKD transfers the knowledge on the target and non-target classes independently. To compare the effectiveness of the knowledge transfer, we used the same small and shallow malware classification models as those in the teacher and the student models in Vanilla KD and MLKD with Self-MCKD.
For all methods, including Vanilla KD, MLKD, and Self-MCKD, we set the temperature factor to 4, following the common practices in knowledge distillation experiments [21,30,36].
As shown in Table 3, we observed that Vanilla KD and MLKD also improved the small and shallow malware classification models. For example, on the MMCC dataset, the VGG8 model with Vanilla KD achieved an accuracy, recall, precision, and F1-score of 98.44%, 98.34%, 98.36%, and 98.29%, respectively, while the VGG8 baseline model achieved 97.70%, 97.70%, 97.67%, and 97.68%, respectively. Also, for ResNet32, MLKD achieved an accuracy, recall, precision, and F1-score of 97.70%, 97.70%, 97.74%, and 97.57%, respectively, while the ResNet32 baseline model achieved an accuracy, recall, precision, and F1-score of 96.87%, 97.42%, 97.39%, and 97.39%, respectively. Similarly, for the Malimg dataset, it was observed that the VGG8 model with Vanilla KD and MLKD improved its malware classification accuracy, recall, precision, and F1-score by approximately 0.3%. Especially, MLKD outperformed Vanilla KD in the recall, precision, and F1-score across both datasets. However, although Vanilla KD and MLKD reduced the training time by approximately 1.5 times compared to the baseline, MLKD required a slightly longer training time than Vanilla KD.
Even though Vanilla KD and MLKD improved the performance of the small and shallow malware classification models, Self-MCKD outperformed both methods in all metrics. For example, on the MMCC dataset, the ResNet32 and VGG8 models with Self-MCKD surpassed the accuracy, recall, precision, and F1-score of the same models with Vanilla KD and MLKD by about 0.2% to 0.8%. Similarly, on the Malimg dataset, the ResNet32 model with Self-MCKD achieved a higher accuracy, recall, precision, and F1-score than these values for the same models with Vanilla KD and MLKD by about 0.4%. Notably, Self-MCKD showed shorter training times compared to those for the baseline and the other knowledge distillation methods.
In Figure 4, we show the ROC curves along with the AUC scores for each method on the MMCC dataset. As shown for class index 8 (e.g., Gatak), indicated by the light green line in Figure 4, Self-MCKD outperformed the other methods, achieving an AUC score exceeding 0.4. In Figure 5, we also visualize the T-SNE results for each method on the Malimg dataset. Compared to the baseline and the other knowledge distillation methods, Self-MCKD produced more distinct and well-separated representations.

6. Conclusions

In this paper, we showed that traditional knowledge distillation in AI-based malware classification methods reduces the effectiveness of knowledge transfer by combining the knowledge of the target malware class and non-target classes. Also, this approach often relies on a high-performing, large, and deep teacher model, which requires high computational costs during the training process, further reducing the efficiency of knowledge transfer. To address these problems, we proposed a novel knowledge distillation technique called Self-MCKD. Its main idea is to transfer two types of knowledge, NMKD and TMKD, by separating the output logits of the target class and the non-target classes. Also, Self-MCKD utilized a pre-trained small and shallow student model to serve as the teacher model. From the experimental results using various malware datasets, we showed that the separation of the output logits in Self-MCKD, consisting of NMKD and TMKD, increased the effectiveness of malware knowledge transfer. Specifically, Self-MCKD outperformed the traditional knowledge distillation in all evaluation metrics. Furthermore, we showed that the small and shallow malware classification methods with Self-MCKD achieved a better performance than the large and deep malware classification methods while requiring only minimal additional computational costs. Also, we showed that Self-MCKD improved the efficiency of malware knowledge transfer by using small and shallow methods as both the teacher and student models. From such experimental results, we believe that Self-MCKD can enhance the malware knowledge in terms of the effectiveness and efficiency of malware classification better than the existing knowledge distillation techniques. In future work, we will validate Self-MCKD under scenarios with highly imbalanced class distributions to assess its robustness and adaptability in real-world malware classification tasks.

Author Contributions

Methodology, data collection, analysis, and interpretation of the results and writing, H.-J.J.; review and editing, H.-J.L., G.-N.K. and S.-H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Republic of Korea, under the ITRC (Information Technology Research Center) support program (IITP-2025-RS-2023-00259967) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation) and a National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. RS-2023-00243075).

Data Availability Statement

MMCC data presented in this study is openly available at https://www.kaggle.com/c/malware-classification, reference number [33]. And Malimg data presented in this study is openly available at https://paperswithcode.com/paper/malware-images-visualization-and-automatic, reference number [32].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kan, Z.; Wang, H.; Xu, G.; Guo, Y.; Chen, X. Towards light-weight deep learning based malware detection. In Proceedings of the 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; Volume 1, pp. 600–609. [Google Scholar]
  2. Moon, J.; Kim, S.; Song, J.; Kim, K. Study on machine learning techniques for malware classification and detection. KSII Trans. Internet Inf. Syst. 2021, 15, 12. [Google Scholar]
  3. Wang, Z.; Li, Z.; He, D.; Chan, S. A lightweight approach for network intrusion detection in industrial cyber-physical systems based on knowledge distillation and deep metric learning. Expert Syst. Appl. 2022, 206, 117671. [Google Scholar] [CrossRef]
  4. Idika, N.; Mathur, A.P. A survey of malware detection techniques. Purdue Univ. 2007, 48, 32–46. [Google Scholar]
  5. You, I.; Yim, K. Malware obfuscation techniques: A brief survey. In Proceedings of the 2010 International Conference on Broadband, Wireless Computing, Communication and Applications, Fukuoka, Japan, 4–6 November 2010; pp. 297–300. [Google Scholar]
  6. Cui, Z.; Xue, F.; Cai, X.; Cao, Y.; Wang, G.; Chen, J. Detection of malicious code variants based on deep learning. IEEE Trans. Ind. Inform. 2018, 14, 3187–3196. [Google Scholar] [CrossRef]
  7. Nataraj, L.; Yegneswaran, V.; Porras, P.; Zhang, J. A comparative assessment of malware classification using binary texture analysis and dynamic analysis. In Proceedings of the 4th ACM Workshop on Security and Artificial Intelligence, Chicago, IL, USA, 21 October 2011; ACM: New York, NY, USA, 2011; pp. 21–30. [Google Scholar]
  8. Hardy, W.; Chen, L.; Hou, S.; Ye, Y.; Li, X. DL4 MD: A deep learning framework for intelligent malware detection. In Proceedings of the International Conference on Data Science (ICDATA); The Steering Committee of The World Congress in Computer Science, Computer Engineering and Applied Computing (WorldComp): Las Vegas, NV, USA, 2016. [Google Scholar]
  9. Drew, J.; Moore, T.; Hahsler, M. Polymorphic malware detection using sequence classification methods. In Proceedings of the 2016 IEEE Security and Privacy Workshops (SPW), San Jose, CA, USA, 22–26 May 2016; pp. 81–87. [Google Scholar]
  10. Ahmadi, M.; Ulyanov, D.; Semenov, S.; Trofimov, M.; Giacinto, G. Novel feature extraction, selection and fusion for effective malware family classification. In Proceedings of the Sixth ACM Conference on Data and Application Security and Privacy, New Orleans, LA, USA, 9–11 March 2016; ACM: New York, NY, USA, 2016; pp. 183–194. [Google Scholar]
  11. Gibert, D. Convolutional Neural Networks for Malware Classification. Master’s Thesis, University Rovira i Virgili, Tarragona, Spain, 2016. [Google Scholar]
  12. Kalash, M.; Rochan, M.; Mohammed, N.; Bruce, N.D.B.; Wang, Y.; Iqbal, F. Malware classification with deep convolutional neural networks. In Proceedings of the 9th IFIP International Conference on New Technologies, Mobility and Security (NTMS), Paris, France, 26–28 February 2018; pp. 1–5. [Google Scholar]
  13. Prajapati, P.; Stamp, M. An empirical analysis of image-based learning techniques for malware classification. In Malware Analysis Using Artificial Intelligence and Deep Learning; Springer: Cham, Switzerland, 2021; pp. 411–435. [Google Scholar]
  14. Fathurrahman, A.; Bejo, A.; Ardiyanto, I. Lightweight convolution neural network for image-based malware classification on embedded systems. In Proceedings of the 2021 International Seminar on Machine Learning, Optimization, and Data Science (ISMODE), Jakarta, Indonesia, 29–30 January 2022; pp. 12–16. [Google Scholar]
  15. Su, J.; Vasconcellos, D.V.; Prasad, S.; Sgandurra, D.; Feng, Y.; Sakurai, K. Lightweight classification of IoT malware based on image recognition. In Proceedings of the 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; Volume 2, pp. 664–669. [Google Scholar]
  16. Jiang, K.; Bai, W.; Zhang, L.; Chen, J.; Pan, Z.; Guo, S. Malicious code detection based on multi-channel image deep learning. J. Comput. Appl. 2021, 41, 1142. [Google Scholar]
  17. Runzheng, W.; Jian, G.; Xin, T.; Mengqi, Y. Research on malicious code family classification combining attention mechanism. J. Front. Comput. Sci. Technol. 2021, 15, 881. [Google Scholar]
  18. Zhu, D.; Xi, T.; Jing, P.; Wu, D.; Xia, Q.; Zhang, Y. A transparent and multimodal malware detection method for android apps. In Proceedings of the 22nd International ACM Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems, Miami Beach, FL, USA, 25–29 November 2019; ACM: New York, NY, USA, 2019; pp. 51–60. [Google Scholar]
  19. Liu, Z.; Sun, M.; Zhou, T.; Huang, G.; Darrell, T. Rethinking the value of network pruning. arXiv 2018, arXiv:1810.05270. [Google Scholar]
  20. Gholami, A.; Kim, S.; Dong, Z.; Yao, Z.; Mahoney, M.W.; Keutzer, K. A survey of quantization methods for efficient neural network inference. In Low-Power Computer Vision; Chapman and Hall/CRC: Boca Raton, FL, USA, 2022; pp. 291–326. [Google Scholar]
  21. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  22. Zhi, Y.; Xi, N.; Liu, Y.; Hui, H. A lightweight android malware detection framework based on knowledge distillation. In Network and System Security: 15th International Conference, NSS 2021, Tianjin, China, 23 October 2021; Springer: Cham, Switzerland, 2021; pp. 116–130. [Google Scholar]
  23. Xia, M.; Xu, Z.; Zhu, H. A Novel Knowledge Distillation Framework with Intermediate Loss for Android Malware Detection. In Proceedings of the IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE), Gold Coast, QLD, Australia, 8–20 December 2022; pp. 1–6. [Google Scholar]
  24. Wang, L.H.; Dai, Q.; Du, T.; Chen, L. Lightweight intrusion detection model based on CNN and knowledge distillation. Appl. Soft Comput. 2024, 165, 112118. [Google Scholar] [CrossRef]
  25. Shen, J.; Yang, W.; Chu, Z.; Fan, J.; Niyato, D.; Lam, K.Y. Effective intrusion detection in heterogeneous internet-of-things networks via ensemble knowledge distillation-based federated learning. arXiv 2024, arXiv:2401.11968. [Google Scholar]
  26. Buciluǎ, C.; Caruana, R.; Niculescu-Mizil, A. Model compression. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 20–23 August 2006; ACM: New York, NY, USA, 2006; pp. 535–541. [Google Scholar]
  27. Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. Fitnets: Hints for thin deep nets. arXiv 2014, arXiv:1412.6550. [Google Scholar]
  28. Yim, J.; Joo, D.; Bae, J.; Kim, J. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4133–4141. [Google Scholar]
  29. Chen, P.; Liu, S.; Zhao, H.; Jia, J. Distilling knowledge via knowledge review. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual Conference, 20–25 June 2021; pp. 5008–5017. [Google Scholar]
  30. Zhao, B.; Cui, Q.; Song, R.; Qiu, Y.; Liang, J. Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11953–11962. [Google Scholar]
  31. Yuan, L.; Tay, F.E.H.; Li, G.; Wang, T.; Feng, J. Revisiting knowledge distillation via label smoothing regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3903–3911. [Google Scholar]
  32. Nataraj, L.; Karthikeyan, S.; Jacob, G.; Manjunath, B.S. Malware images: Visualization and automatic classification. In Proceedings of the 8th International Symposium on Visualization for Cyber Security, Pittsburgh, PA, USA, 20 July 2011; pp. 1–7. [Google Scholar]
  33. Ronen, R.; Radu, M.; Feuerstein, C.; Yom-Tov, E.; Ahmadi, M. Microsoft malware classification challenge. arXiv 2018, arXiv:1802.10135. [Google Scholar]
  34. Huang, D.; Li, T.O.; Xie, X.; Cui, H. Themis: Automatic and Efficient Deep Learning System Testing with Strong Fault Detection Capability. In Proceedings of the 2024 IEEE 35th International Symposium on Software Reliability Engineering (ISSRE), Tsukuba, Japan, 28–31 October 2024; pp. 451–462. [Google Scholar]
  35. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  36. Jin, Y.; Wang, J.; Lin, D. Multi-level logit distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 17–24 June 2023; pp. 24276–24285. [Google Scholar]
Figure 1. Difference between traditional knowledge distillation and proposed Self-MCKD.
Figure 1. Difference between traditional knowledge distillation and proposed Self-MCKD.
Electronics 14 01077 g001
Figure 2. The differences between feature-based (a) and logit-based (b) distillation approaches.
Figure 2. The differences between feature-based (a) and logit-based (b) distillation approaches.
Electronics 14 01077 g002
Figure 3. Structure of Self-MCKD.
Figure 3. Structure of Self-MCKD.
Electronics 14 01077 g003
Figure 4. Comparison of ROC curves for knowledge distillation methods under the MMCC dataset.
Figure 4. Comparison of ROC curves for knowledge distillation methods under the MMCC dataset.
Electronics 14 01077 g004
Figure 5. Comparison of T-SNE visualizations for knowledge distillation methods under the Malimg dataset.
Figure 5. Comparison of T-SNE visualizations for knowledge distillation methods under the Malimg dataset.
Electronics 14 01077 g005
Table 1. Performance variation with changes in α and β factors of Self-MCKD.
Table 1. Performance variation with changes in α and β factors of Self-MCKD.
MMCC( α , β )(0.1, 0.9)(0.2, 0.8)(0.3, 0.7)(0.4, 0.6)(0.5, 0.5)
Accuracy98.25%98.07%98.44%98.34%98.52%
( α , β )(0.6, 0.4)(0.7, 0.3)(0.8, 0.2)(0.9, 0.1) p n t m t S
Accuracy98.25%98.62%98.44%98.25%98.44%
Malimg( α , β )(0.1, 0.9)(0.2, 0.8)(0.3, 0.7)(0.4, 0.6)(0.5, 0.5)
Accuracy99.25%99.25%99.36%99.25%99.46%
( α , β )(0.6, 0.4)(0.7, 0.3)(0.8, 0.2)(0.9, 0.1) p n t m t S
Accuracy99.57%99.57%99.57%99.57%99.36%
Bold values indicate the highest accuracy observed for each dataset.
Table 2. Comparison results on computational complexity and performance with baseline models under various malware datasets.
Table 2. Comparison results on computational complexity and performance with baseline models under various malware datasets.
DatasetModelFLOPsMACsParamsAccuracyTraining Time
MMCCBaselineVGG16627.04 M313.52 M14.77 M98.25%91 m 8 s
VGG8136.01 M68.00 M3.96 M97.70%42 m 23 s
ResNet110508.42 M254.21 M1.74 M97.79%94 m 17 s
ResNet32138.89 M69.45 M0.47 M96.87%48 m 39 s
Self-MCKD + VGG8272.02 M136.01 M3.96 M98.62%28 m 19 s
Self-MCKD + ResNet32277.78 M138.90 M0.47 M98.53%35 m 21 s
MalimgBaselineVGG16627.04 M313.52 M14.77 M99.25%72 m 52 s
VGG8136.01 M68.00 M3.96 M99.04%33 m 18 s
ResNet110508.42 M254.21 M1.74 M99.36%82 m 5 s
ResNet32138.89 M69.45 M0.47 M99.04%41 m 51 s
Self-MCKD + VGG8272.02 M136.01 M3.96 M99.57%27 m 11 s
Self-MCKD + ResNet32277.78 M138.90 M0.47 M99.25%30 m 52 s
Bold values indicate the highest accuracy or the shortest training time observed among models of the same type for each dataset.
Table 3. Comparison of results for knowledge distillation methods using various malware datasets.
Table 3. Comparison of results for knowledge distillation methods using various malware datasets.
DatasetBaselineTechniqueAccuracyRecallPrecisionF1-ScoreTraining Time
MMCCVGG8None97.70%97.70%97.67%97.68%42 m 23 s
Vanilla KD98.44%98.34%98.36%98.29%28 m 43 s
MLKD98.43%98.44%98.44%98.37%33 m 34 s
Self-MCKD98.62%98.71%98.73%98.69%28 m 19 s
ResNet32None96.87%97.42%97.39%97.39%48 m 39 s
Vanilla KD97.79%97.06%96.83%96.82%35 m 26 s
MLKD97.70%97.70%97.74%97.57%38 m 20 s
Self-MCKD98.53%98.44%98.47%98.41%35 m 21 s
MalimgVGG8None99.04%98.72%98.81%98.73%33 m 18 s
Vanilla KD99.36%99.04%99.08%99.03%27 m 13 s
MLKD99.25%99.25%99.32%99.25%29 m 1 s
Self-MCKD99.57%99.25%99.29%99.25%27 m 11 s
ResNet32None99.04%99.25%99.28%99.26%41 m 51 s
Vanilla KD99.14%98.82%98.96%98.80%30 m 54 s
MLKD98.92%98.82%98.85%98.82%31 m 39 s
Self-MCKD99.25%99.25%99.36%99.26%30 m 52 s
Bold values indicate the highest Accuracy, Recall, Precision, or F1-score, or the shortest training time observed among models for each dataset.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jeong, H.-J.; Lee, H.-J.; Kim, G.-N.; Choi, S.-H. Self-MCKD: Enhancing the Effectiveness and Efficiency of Knowledge Transfer in Malware Classification. Electronics 2025, 14, 1077. https://doi.org/10.3390/electronics14061077

AMA Style

Jeong H-J, Lee H-J, Kim G-N, Choi S-H. Self-MCKD: Enhancing the Effectiveness and Efficiency of Knowledge Transfer in Malware Classification. Electronics. 2025; 14(6):1077. https://doi.org/10.3390/electronics14061077

Chicago/Turabian Style

Jeong, Hyeon-Jin, Han-Jin Lee, Gwang-Nam Kim, and Seok-Hwan Choi. 2025. "Self-MCKD: Enhancing the Effectiveness and Efficiency of Knowledge Transfer in Malware Classification" Electronics 14, no. 6: 1077. https://doi.org/10.3390/electronics14061077

APA Style

Jeong, H.-J., Lee, H.-J., Kim, G.-N., & Choi, S.-H. (2025). Self-MCKD: Enhancing the Effectiveness and Efficiency of Knowledge Transfer in Malware Classification. Electronics, 14(6), 1077. https://doi.org/10.3390/electronics14061077

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop