Next Article in Journal
A Temporal Window Attention-Based Window-Dependent Long Short-Term Memory Network for Multivariate Time Series Prediction
Previous Article in Journal
Optimization of Two-Phase Ejector Mixing Chamber Length under Varied Liquid Volume Fraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cluster-Based Structural Redundancy Identification for Neural Network Compression

1
State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
2
Key Laboratory of Networked Control Systems, Chinese Academy of Sciences, Shenyang 110016, China
3
Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110169, China
4
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 9; https://doi.org/10.3390/e25010009
Submission received: 6 October 2022 / Revised: 15 December 2022 / Accepted: 15 December 2022 / Published: 21 December 2022

Abstract

:
The increasingly large structure of neural networks makes it difficult to deploy on edge devices with limited computing resources. Network pruning has become one of the most successful model compression methods in recent years. Existing works typically compress models based on importance, removing unimportant filters. This paper reconsiders model pruning from the perspective of structural redundancy, claiming that identifying functionally similar filters plays a more important role, and proposes a model pruning framework for clustering-based redundancy identification. First, we perform cluster analysis on the filters of each layer to generate similar sets with different functions. We then propose a criterion for identifying redundant filters within similar sets. Finally, we propose a pruning scheme that automatically determines the pruning rate of each layer. Extensive experiments on various benchmark network architectures and datasets demonstrate the effectiveness of our proposed framework.

1. Introduction

Compared with traditional machine learning methods, deep learning-based methods have greatly improved the performance of many computing tasks, such as image recognition [1], object detection [2] and speech segmentation [3]. Edge AI [4,5,6] stands out as a disruptive technology for 6G by embedding model training and inference capabilities at the edge of the network, which seamlessly integrates perception, communication, computing and intelligence to improve the efficiency, effectiveness, privacy and security of 6G networks. However, in pursuit of better performance, it is often at the expense of increasing computing power. It has gradually increased from the earliest LeNet [7] to approximately 20 layers of VGG [8]; in particular, the commonly used ResNet [9] and DenseNet [10] have increased astonishingly to hundreds of layers. The over-parameterization and redundant computation of the network makes it difficult to deploy on edge devices with limited computing resources. At present, model compression techniques for how to obtain a more efficient network have been proposed successively, including knowledge distillation [11], quantization [12,13], low-rank decomposition [14] and network pruning [15,16,17,18]. Among them, network pruning, which can dynamically evolve the baseline network into a more efficient sub-network, has become a widely recognized model compression method.
Network pruning is mainly divided into unstructured pruning and structured pruning. Unstructured pruning prunes individual weights in the model to compress the DNN (Deep Neural Network) [19]. However, although this method can greatly reduce the parameters, the generated unstructured matrices require a special sparse matrix operation library, which limits its practical acceleration in the general CNN acceleration framework. Structured pruning uses convolution kernels/channels or layers as pruning granularity for pruning. It has received widespread attention because of its advantages of being directly compatible with current general-purpose hardware and highly efficient basic linear algebra subprogram (BLAS) libraries. The research in this paper belongs to the category of structured pruning.
Typical filter pruning includes three stages [20]: (1) training a large, over-parameterized model (sometimes a pre-trained model), (2) pruning the trained large model according to certain criteria, and (3) fine-tuning the pruned model to restore the lost performance. Although existing pruning methods have achieved good results, there are still many problems. To assess the importance of filters, recently, a variety of filter pruning methods have been proposed to design more effective pruning guidelines, such as the average percentage of zero values (APoZ) [17], L 1 -norm [18], Taylor expansion [21], sparsity norm [22], geometric median (FPGM) [23], high rank (Hrank) [24] and variants of the pruning mask [25]. Due to the different distributions of the values of the convolution kernels in different layers, the abovementioned pruning methods based on global or local criteria for sorting filters may ignore filters with smaller values in the sorting but extract edge features. Huang et al. [26] compared different pruning standards and found that they have strong similarities, and that the importance of the obtained filters is almost the same, resulting in similar pruning structures. Recent work shows that the pruning structure is the key to determining the performance of the pruning model rather than the inheritance weight. Manually setting the pruning rate of each convolutional layer is equivalent to redesigning the network structure completely, and improper pruning rate settings will result in insufficient pruning or excessive pruning. In addition, for large networks, it is very expensive to accurately calculate the importance of the filters and set the pruning rate of each layer.
In this paper, we propose a clustering-based dynamic pruning method considering the similarity between filters. Compared with existing importance-based methods, we analyze the relevant information on the representational power among all filters within a layer and remove filters with overlapping functions. The proposed scheme takes into account edge features that are ignored based on importance ranking. Specifically, we cluster all filters within a layer in units of filters and select one deletion in each group whose features can be replaced by other filters. Each layer automatically generates groups according to the parameter distribution to determine the pruning rate of each layer, avoiding the problem of manually specifying the pruning rate. To assess the similarity of the representational power of filters in each group, a criterion is defined to measure the relative importance of all filters within a group. This criterion is scoped to a group and avoids the problem of threshold specification for global or local pruning. Extensive experiments demonstrate that our proposed method is more general than importance-based methods. Figure 1 shows a graphical illustration of our motivation and pruning framework. To summarize, our main contributions are as follows:
  • We propose a novel pruning scheme that does not depend on importance but is based on the similarity between filters for channel-level pruning.
  • We introduce an effective method for measuring the relative importance of filters, avoiding the problems of over-pruning and under-pruning caused by threshold specification.
  • The proposed scheme automatically determines each layer’s pruning rate according to each layer’s parameter distribution, which avoids the problem of unreasonable pruning structure caused by manually specifying the pruning rate.
  • A large number of experiments prove the effectiveness of the algorithm proposed in this paper.

2. Related Work

2.1. Unstructured Pruning

The research on network pruning originated from the 1989 paper on skeletonization [27]. LeCun [16] in 1990 and Hassibi [17] in 1993 proposed OBD and OBS methods, respectively, which measure the importance of the weight in the network based on the second derivative of the loss function relative to the weight (Hessian matrix). Hang Song et al. published a series of works on model compression for deep neural networks [28,29]. Among them, [29] compressed the classical networks AlexNet and VGG at that time, combined with various methods, such as pruning, quantization and Huffman coding to compress the network size by dozens of times. However, unstructured weight pruning does not guarantee GPU acceleration.

2.2. Structured Pruning

To overcome the above limitations, SSL [18] proposed to regularize the structures (i.e., filters, channels, filter shapes and layer depth) of DNNs. This was the first work to actually measure GPU acceleration, and structured pruning gradually became the focus of pruning research. Studies on structured pruning have been proposed one after another. The simplest of them is the magnitude-based weight pruning, which evaluates the importance according to the absolute value of the parameter or feature output [18,22,30,31]. Some studies [22,32,33] considered the impact of pruning on model loss as a criterion for measuring the importance of parameters. For example, Molchanov [21] was also based on Taylor expansion, using the first-order term’s absolute value in the objective function’s expansion relative to the activation function as pruning criteria. In addition, [34,35] considered the effect of pruning on the re-constructability of feature output, which minimizes the reconstruction error of the pruned network for feature output.
There are also other criteria based on the weights of the importance of ranking. He et al. [23] proposed a filter pruning via the geometric median (FPGM) method, the basic idea of which was to remove redundant parameters based on the geometric median. Lin et al. [24] developed a mathematically formulated method to prune filters with low-rank feature maps. The disadvantage of the above greedy algorithm is that it can only find the optimal local solution, which ignores the relationship between parameters. Some studies [36,37,38] try to consider the relationship between parameters, trying to find a better global solution. For example, Peng [38] proposed the collaborative channel pruning (CCP) method, which considers the dependencies between channels, formalizes the channel selection problem as a quadratic programming problem under constraints, and then uses sequential quadratic programming to solve it.

2.3. Other Compression Techniques

Other types of DNN model compression techniques are also being explored. Quantization [39,40] compressed the model by reducing the size of the weights or activations. XNOR-Net [41] and BinaryNet [42] used binary weights and activations to compress the model; [43,44] studied how to choose the appropriate quantization parameters to minimize the impact on the accuracy as much as possible; [45,46] explored how to make the distribution of quantified objects more suitable for quantification, and [47] introduced quantization operations during training to explore more efficient training of low-precision quantization networks. The quantization of ultra-low precision [48] and hybrid precision [49] has also been a popular topic in recent years. Knowledge distillation trains another simple network by using the output of a pre-trained complex network as a supervisory signal. The studies in [50,51] improved the prediction performance of the student model by adjusting the temperature; [52] transferred the knowledge of multiple teachers to a single student model so that the trained student model could handle the original tasks of the multiple teacher models, and [53] used different knowledge forms including output feature, intermediate feature, relational feature and structural feature. The low-rank decomposition sparse convolution kernel matrix was created by merging dimensions and imposing low-rank constraints. Since most of the weight vectors are distributed in the low-rank subspace, the convolution kernel matrix can be reconstructed with a few basis vectors to reduce the storage space. Jaderberg et al. [54] decomposed the convolution kernel of w × h into w × 1 and 1 × h , and reconstructed the learned dictionary weight linearly to obtain the output feature map; Liu et al. [8] used a two-stage decomposition method to study the redundancy between the channels; Wang et al. [55] proposed fixed-point decomposition, and then restored the performance through pseudo-full-precision weight repetition, weight balancing and fine-tuning; Kim et al. [56] proposed tucker decomposition, which performs binary decomposition of the first tensor along the input channel dimension to obtain convolutions of w × 1 , 1 × h and 1 × 1 , and Lebedev et al. [57] proposed CP decomposition on the basis of ternary decomposition.
The current model compression methods through pruning still determine the importance of a single parameter or filter by looking for a criterion and combining the pruning method and processing to restore the performance of the pruned model. Different from previous methods, we use clustering to find filters with overlapping functions more effectively by comparing the similarity between filters.

3. Methodology

In this section, we propose a novel pruning scheme, which is the cluster similarity-based filter pruning method. We first introduce the overall framework and related notation definitions, and then describe the motivation and implementation details. Finally, we propose the corresponding pruning scheme.

3.1. Overall Framework

We show the flow of a single pruning of the proposed pruning scheme in Figure 1. First, we use a filter as the unit to cluster all the convolution kernels in one layer and obtain the cluster label corresponding to each filter. We improved the single-threshold selection from the previous single-dimensional space to map the filter into the multi-dimensional space through clustering. The clustering results are used to guide the determination of redundant filters in the next step. Since filters with similar features are in the same cluster set, we only need to find redundant filters in each cluster space. Then, redundant filters in each cluster are obtained according to the proposed criterion and removed. Finally, after several iterations, we can get the final compact network structure without specifying the pruning rate of each layer and reach the specified pruning rate constraint.

3.2. Motivation and Definitions

One of the main problems of filter pruning is how to select effective filters and retain as much of the expressive power of the original network as possible. Current amplitude-based pruning algorithms all rely on the assumption that removing relatively insignificant weights in the network has little effect on the pruned network performance. Unlike current views based on parameter importance, we propose that the removal of any one of the filters will not significantly impair the representational power of the network as long as there are two sufficiently similar channels. This reduction also resonates with the well-known Hebbian principle, which roughly states that “neurons which fire together, wire together”. The visualization results of the filters and feature maps of the first convolutional layer of VGG16 are shown in Figure 2. It can be seen that there are a large number of similar filters in the trained network.
Assuming that the neural network has L convolutional layers, N l and N l + 1 represent the number of input channels and output channels of the l t h layer convolution layer, respectively. F ( l , i ) represents the i t h filter of the l t h layer, and the corresponding input feature map can be expressed as F ( l , i ) R H × W × B , where H , W , B represent the height and width of the feature maps and the batch size, respectively. The tensor of the connections of the l t h and l + 1 t h layers can be parameterized by W R N l × N l + 1 × K × K , 1 l L .
We demonstrate that if there is a set of similar filters within a layer, pruning filters randomly or selectively in that set is better than pruning the least important filters within that layer. We assume that there is a set of similar filters S α and another S β with remaining filters in the l t h layer, containing n and m filters, respectively. We choose positive constants a , b > 0 and use the random events ( i = 1 n α i a ) and ( i = 1 m β i b ) to indicate that the filters in S α and S β perform better, and use the sum of the two to indicate the performance of the entire layer. When removing a filter from the l t h layer, there are several situations including (1) no pruning; (2) randomly selecting a filter to prune in S α ; (3) pruning according to the minimum rule in S α ; (4) pruning according to the minimum rule in S β ; (5) pruning the least important filter in the layer, i.e., min ( S α , S β ) :
p o = P ( i = 1 n α i a ) + P ( i = 1 m β i b )
p α r = P ( i = 1 n 1 α i a ) + P ( i = 1 m β i b )
p α _ = P ( i = 1 n α i α _ a ) + P ( i = 1 m β i b )
p β _ = P ( i = 1 n α i a ) + P ( i = 1 m β i β _ b )
p g = n m + n p α _ + m m + n p β _
Note that 0 α n α _ α n ; therefore, we have
P ( i = 1 n 1 α i a ) P ( i = 1 n α i α _ a ) P ( i = 1 n α i a )
indicating that p α r p α _ p o . For any filter, the contribution to the network cannot be infinite, where the variance is uniformly bounded.
C 1 > 0 , s . t . D η i C 1 , i = 1 , 2 , , n
By Chebyshev’s inequality, for any real number ϵ > 0 ,
P ( 1 n i = 1 n α i E α i ϵ ) D ( i = 1 n α i ) ϵ 2 n 2
From Equation (7) we can get:
Cov α i , α j D α i D α j C 1
We assume that there are C 2 n   ( 0 C 2 1 ) pairs of similar filters in the set S α , i.e., # { ( i , j ) : Cov α i , α j 0 , i j , i , j = 1 , , n . C 2 n , then we have:
D ( i = 1 n α i ) = i = 1 n D α i + i j Cov α i , α j C 1 n + C 1 C 2 n = C 1 1 + C 2 n
Available by Equation (8):
P ( 1 n i = 1 n α i E α i ϵ ) C 1 1 + C 2 ϵ 2 n 0
This means 1 n i = 1 n α i E α i converges in probability to zero, i.e., 1 n i = 1 n α i E α i P 0 . We consider the filter’s contribution to be a positive number that expects a uniform positive lower bound:
ϵ 0 > 0 ,   s . t . E α i ϵ 0 , i = 1 , 2 , , n
By Equation (9) we can get:
P ( 1 n i = 1 n ( α i E α i ) > ϵ 0 2 ) = P ( i = 1 n α i > i = 1 n E α i ϵ 0 2 n ) = P ( i = 1 n α i > ϵ 0 2 n + i = 1 n ( E α i ϵ 0 ) ) P ( i = 1 n α i > ϵ 0 2 n ) P ( i = 1 n α i > b )
Letting n + , then 1 n i = 1 n α i E α i P 0 , and we have:
lim n P ( i = 1 n α i > b ) lim n P ( 1 n i = 1 n α i E α i > ϵ 0 2 ) = 1
lim n P ( i = 1 n α i α r > b ) = lim n P ( i = 1 n α i α _ > b ) = 1
Therefore, we have p α r p α _ p o for an n that is large enough. Furthermore, p β _ p o p α _ , and since p g is the average of p α _ and p β _ , there is p β _ p g p α _ . In summary, we have p β _ p g p α r p α _ p o , which indicates that pruning filters in a similar set (even randomly) works better than pruning the least important filter in that layer. This provides the basis for our pruning below. The next challenge is how to find similar sets and how to determine which filters within a set are redundant.

3.3. Cluster Pruning

K-means clustering provides a solution for how to find similar convolution kernels within a layer. For the l t h layer, N l + 1 filters are divided into k clusters, then the clustering result is { S 1 , S 2 , S 3 , , S k } , k N l + 1 , where S 1 = { F 1 , F 3 , F 7 , F 9 , F 12 } , S 2 = { F 2 , F 5 , F 14 } ,…, S k = { F 6 , F 10 , F 13 , F N l + 1 } . We divide each layer into different sets in which convolutional kernels contain similar feature information. At the same time, we also get the centroid { C 1 , C 2 , C 3 , , C k } , representing the set information:
C k = F i S k F i S k
where C k represents the center of the k t h cluster, S k represents the number of objects in the k t h cluster and F i represents the i t h object in the k t h cluster.
After getting sets containing similar information, the task is to find redundant filters in each set. The cluster center anchors the representative information of all filters in a similar cluster. After obtaining the cluster center of each similar cluster, it is intuitive to choose to keep the cluster center and delete other filters of the same cluster. However, this simple pruning method causes a large number of filters to be discarded, resulting in a sharp drop in performance. We look for a reasonable criterion for progressive pruning to determine redundant kernels in a cluster. In contrast to the intuitive pruning approach, we propose that cluster centers can be pruned. Since the cluster center is the mean value of all filters in clusters in each dimension, the cluster center of the remaining filters is still in the original cluster center position after removing the cluster center. This indicates that the information contained in all filters in the cluster is not lost after removing the cluster center.
The above claim is in the ideal state, that is, the centroid of each cluster is exactly one element in the set. In experiments, it is difficult to find the element in the cluster set that happens to be the cluster center for each clustering. We find a compromise scheme, where each time we find the element closest to the cluster center, the information represented by this element can be transferred to other filters, and the original cluster set still retains the original information. First, we calculate the distance between all filters and cluster center in the k t h cluster set of the l t h layer:
D i k = d i s t ( F i k , C k ) , 1 i S k = n = 1 N l + 1 k 1 = 1 K k 2 = 1 K W i n , k 1 , k 2 C k ( n , k 1 , k 2 ) 2
where W i n , k 1 , k 2 is each weight parameter in the i t h filter, C k ( n , k 1 , k 2 ) is each weight parameter in the cluster center of the k t h cluster and F i k is the i t h filter in the k t h cluster. Then, the filter that needs to be pruned in the similar cluster is:
P k = arg   min   1 i S k D i k
where P k is the filter to be pruned in the k t h cluster. Finally, the set of all pruned filters in the l t h layer is { P 1 l , P 2 l , P 3 l , , P k l } .

3.4. Pruning Scheme

In the above section, we propose how to identify redundant filters from a clustering set. However, for the number of hundreds or even thousands of filters in the current network structure, the number of redundant filters obtained by one clustering is far from enough. Therefore, we propose an iterative pruning scheme to identify more redundant filters to meet the compression requirements of large pruning rates and large network structures. After one pruning is completed, the next pruning continues to cluster to find a new similarity set and delete the redundant filters in the set. Compared with the current method of manually specifying the pruning rate of each layer, we only need to adjust one parameter to control the pruning rate of all layers. This is very efficient for large networks with hundreds of layers.
The overall workflow of our cluster-based pruning algorithm is shown in Algorithm 1 and can be summarized as follows:
  • For each convolutional layer, first initialize each cluster center { μ 1 , μ 2 , μ 3 , , μ k } and compare any filter in the layer with each cluster center to construct a N l + 1 × k distance matrix. In each iteration, λ i = arg min k { 1 , 2 , 3 , , k } d i , k is obtained in each row, and the corresponding filter is divided into the corresponding cluster k , and, finally, the cluster set S l = { S 1 l , S 2 l , S 3 l , , S k l } and cluster center C l = { C 1 l , C 2 l , C 3 l , , C k l } are obtained.
  • For each cluster set S k l obtained, each filter in the set and the cluster center C k l obtain a S k l -dimensional vector according to Equation (2), and the filter corresponding to the minimum value in the vector is determined as the layer that needs to be pruned filter.
  • After one pruning, calculate the pruning end condition, that is, the amount of computation r a t e F L O P s or parameters r a t e p a r a m s after pruning, prune in a loop until the given pruning rate is reached and fine-tune the generated model to restore performance.
Algorithm 1: Iterative pruning algorithm.
Input: Training dataset D ; the model with W , and each layer with W R N l × N l + 1 × K × K , 1 l L ; FLOPs or params pruning rate: r a t e F L O P s / r a t e p a r a m s .
Output: The pruned model W P
1:  W train ( W , D )
2:  while pruned rate = 0 to rate do
3:      for   l = 1 , 2 , , L  do
4:            initialize the clusters { μ 1 , μ 2 , μ 3 , , μ k }
5:           for i = 1 , 2 , , N l + 1  do
6:                 d i , k = F i μ k 2
7:                 λ i = arg min k { 1 , 2 , 3 , , k } d i , k
8:                 S λ i = S λ i { F i }
9:            end for
10:            S l = { S 1 l , S 2 l , S 3 l , , S k l }
11:            C l = { C 1 l , C 2 l , C 3 l , , C k l }
12:            for  k = 1 , 2 , , k  do
13:                 for j = 1 , 2 , , S k l  do 
14:                      D j k = d i s t ( F j k , C k l )
15:                      P k = arg min 1 i S k l D j k
16:                 end for
17:            end for
18:          P l = { P 1 l , P 2 l , P 3 l , , P k l }
19:         end for 
20:        W P W P
21:  end while
22:  W f i n e t u n e ( W P , D )
In the network structure, the processing of special structures, such as dense block, residual block and inverted residual block, also has a greater impact on the final compression performance. In these structures, we process in blocks and perform pruning on the basis of satisfying the original relative relationship. For example, in a residual block, the number of input channels and output channels of each block is the same, and there are also constraints on the number of three convolutional layers in the block. According to previous work [9], the 3 × 3 convolution kernel in the residual block has the same number of input and output channels, that is, the output channels of the previous layer and the input channels of the last layer are the same, as shown in Figure 3.

4. Experiments

4.1. Experimental Settings

We evaluate the effectiveness of our algorithm on CIFAR-10, CIFAR-100 [58] and ILSVRC-2012 [1] datasets using representative CNN architectures VGGNet [8] and ResNet [9]. CIFAR10 contains 50,000 training images and 10,000 testing images (size 32 × 32), which are categorized into 10 different classes. CIFAR100 is similar to CIFAR-10 but has 100 classes. ImageNet contains 1.28 million training images and 50 k validation images of 1000 classes. VGGNet and ResNet represent two typical network structures with single branch and multiple branches, respectively. All experiments are implemented on four NVIDIA TITAN Xp GPUs using PyTorch.
We measure the complexity of the network using floating point operations (FLOPs) required for forward propagation. The computational cost of one convolutional layer is:
F L O P s = H W C i n   K 2 + 1 C o u t
P a r a m s = C i n   K 2 + 1 C o u t
where H and W are the height and width of the input feature map of the layer, respectively, and C i n and C o u t   are the number of input channels and output channels. In this paper, we use the drop rate of FLOPs to evaluate the compression performance of each algorithm, that is, the smaller the accuracy drop of the compressed network model under the same compression ratio, the better the algorithm performance:
r a t e F L O P s = 1 F L O P s o r i g i n a l F L O P s c o m p r e s s e d  
r a t e p a r a m s = 1 P a r a m s o r i g i n a l P a r a m s c o m p r e s s e d  

4.2. Results on CIFAR-10/100 Datasets

We evaluate the effectiveness of the proposed framework using VGG16 [8] and ResNet-32/56/110 [9] on CIFAR10 and CIFAR100 datasets [58] and compare with existing algorithms, such as L1 [18], Molchanov et al. [21], SFP [59], FPGM [23], Hrank [24] and SRR-GR [60]. All the networks are trained using SGD with Nesterov momentum [61] of 0.9, a weight decay parameter of 10 4 and an initial learning rate of 0.1. The learning rate is set to 0.001 when updating parameters or fine-tuning. For VGG16, the baseline network is trained for 300 epochs with a batch size of 256. For ResNet, the baseline network is trained for 200 epochs with a batch size of 256.
It can be seen from Table 1 that our proposed pruning framework achieves less accuracy loss with higher computational compression using VGG16 on the CIFAR10 and CIFAR100 datasets. We find effective channels in the network and reduce the false deletion of channels to achieve better performance than other algorithms. Comparing the results on the CIFAR10 and CIFAR100 datasets, the same network has different redundancy on different datasets, as shown in Figure 4. Due to the large redundancy of VGG16 on the CIFAR10 dataset, our performance differs little from other algorithms when the compression ratio is small. However, when the compression ratio becomes larger, the performance of each algorithm is significantly different. For example, when the compression ratio reaches 90% on CIFAR10, the performance of our algorithm and Hrank is quite different, that is, we identify redundant channels more effectively. However, since VGG16 has less redundancy on CIFAR100, pruning is more difficult. When the pruning rate is approximately 50%, the network performance loss after pruning is obvious; however, we still maintain good performance at larger compression ratios.
The results of ResNet with different depths on the CIFAR10 and CIFAR100 datasets are shown in Table 2. Pruning is more challenging due to the residual structure in ResNet. In addition to the influence of redundant judgment criteria, the processing of residual structures also affects the pruning performance of ResNet. As can be seen from the table, we reduce the accuracy loss while reaching the same or higher compression ratio in ResNet of different depths. Due to the overfitting of the network on the CIFAR10 dataset, the accuracy of pruned ResNets of different depths does not decrease but increases after compression. For example, our framework compresses 60.1% of the FLOPs on the ResNet-56, but the accuracy increased by 0.54%. The uncompressed ResNet performs generally on the CIFAR100 dataset, our algorithm still maintains the original performance after compressing half of the parameters. However, interestingly, the redundancy of ResNet on the CIFAR100 dataset does not increase with network depth, for example, ResNet-110 still has a larger accuracy loss than ResNet-56 with less pruning rate.

4.3. Results on ILSVRC-2012

In the experiments, we use ResNet-18/34/50 to demonstrate the proposed pruning performance on a large-scale dataset, ILSVRC-2012. All the baseline networks are obtained by training 100 epochs with a batch size of 256. We follow the same parameter settings as [20,60]. We compare the proposed method with ThiNet [34], FPGM [23], MIL [62], L1 [18], CP [35], SFP [59], Hrank [24], PGMPF [25] and SRR-GR [60]. All the results of the other methods in the table are directly from their reports in the literature.
The results of ResNet with different depths on ILSVRC-2012 are shown in Table 3. It can be seen from the table that our framework still has good compression performance on large datasets. We reduce the accuracy loss at the same computational compression ratio. Combined with the analysis of the above CIFAR100 dataset, it shows that the network is more sensitive to compression on larger datasets that are underfitting. When the compression ratio increases, the accuracy drops significantly, and the performance of each algorithm varies significantly. For example, on ResNet-18, when other algorithms compress less than half of the calculation, the Top-1 accuracy drops by between 2 and 4%, but ours controls the accuracy loss to within 2%. However, it is common sense that the deeper the network, the greater the redundancy of the model. Under the same compression rate, the accuracy of ResNet-50 only drops by 0.35%, and ResNet-18 drops by 1.82%.

5. Conclusions

Aiming at the problems of ignoring edge features and manually specifying the pruning rate in current importance-based model pruning algorithms, this paper proposes a new model pruning framework based on similarity clustering. We reconsider the redundancy of neural network models from the perspective of similarity and find similar sets of filters through clustering and then propose a criterion for determining filter redundancy in similar sets. In order to solve the problem of a long compression period caused by excessive fine-tuning, we propose a corresponding iterative pruning scheme. Extensive experiments demonstrate the effectiveness of our proposed compression framework, while we cluster filters into multidimensional spaces and reconsider filter redundancy from a similar perspective, without specifying pruning rate. However, multi-dimensional values for clustering are still an unsolved problem. In the next step, we will combine this research with reinforcement learning and continue to mine redundant parameters in the network.

Author Contributions

Validation, C.X.; Writing—original draft, T.W.; Writing—review & editing, C.S. and P.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China under Grant 2017YFA0700300, National Natural Science Foundation of China (61903356), Liaoning Natural Science Foundation (2021-MS-030, 2022JH6/100100013), and Independent project of State Key Laboratory of Robotics 2022-Z03.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
  2. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  3. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  4. Song, C.; Liu, S.; Han, G.; Zeng, P.; Yu, H.; Zheng, Q. Edge intelligence based condition monitoring of beam pumping units under heavy noise in the industrial internet of things for industry 4.0. IEEE Internet Things J. 2022. [Google Scholar] [CrossRef]
  5. Song, C.; Xu, W.; Han, G.; Zeng, P.; Wang, Z.; Yu, S. A cloud edge collaborative intelligence method of insulator string defect detection for power iiot. IEEE Internet Things J. 2020, 8, 7510–7520. [Google Scholar] [CrossRef]
  6. Song, C.; Sun, Y.; Han, G.; Rodrigues, J.J. Intrusion detection based on hybrid classifiers for smart grid. Comput. Electr. Eng. 2021, 93, 107212. [Google Scholar] [CrossRef]
  7. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  8. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  10. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  11. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  12. Courbariaux, M.; Bengio, Y.; David, J.-P. Binaryconnect: Training deep neural networks with binary weights during propagations. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  13. Hubara, I.; Courbariaux, M.; Soudry, D.; El-Yaniv, R.; Bengio, Y. Quantized neural networks: Training neural networks with low precision weights and activations. J. Mach. Learn. Res. 2017, 18, 6869–6898. [Google Scholar]
  14. Denton, E.L.; Zaremba, W.; Bruna, J.; LeCun, Y.; Fergus, R. Exploiting linear structure within convolutional networks for efficient evaluation. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  15. LeCun, Y.; Denker, J.; Solla, S. Optimal brain damage. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 27–30 November 1989; Volume 2. [Google Scholar]
  16. Hassibi, B.; Stork, D. Second order derivatives for network pruning: Optimal brain surgeon. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 30 November–3 December 1992; Volume 5. [Google Scholar]
  17. Hu, H.; Peng, R.; Tai, Y.-W.; Tang, C.-K. Network trimming: A datadriven neuron pruning approach towards efficient deep architectures. arXiv 2016, arXiv:1607.03250. [Google Scholar]
  18. Li, H.; Kadav, A.; Durdanovic, I.; Samet, H.; Graf, H.P. Pruning filters for efficient convnets. arXiv 2016, arXiv:1608.08710. [Google Scholar]
  19. Hinton, G.E.; Osindero, S.; Teh, Y.-W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  20. Liu, Z.; Sun, M.; Zhou, T.; Huang, G.; Darrell, T. Rethinking the value of network pruning. arXiv 2018, arXiv:1810.05270. [Google Scholar]
  21. Molchanov, P.; Tyree, S.; Karras, T.; Aila, T.; Kautz, J. Pruning convolutional neural networks for resource efficient inference. arXiv 2016, arXiv:1611.06440. [Google Scholar]
  22. Liu, Z.; Li, J.; Shen, Z.; Huang, G.; Yan, S.; Zhang, C. Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2736–2744. [Google Scholar]
  23. He, Y.; Liu, P.; Wang, Z.; Hu, Z.; Yang, Y. Filter pruning via geometric median for deep convolutional neural networks acceleration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4340–4349. [Google Scholar]
  24. Lin, M.; Ji, R.; Wang, Y.; Zhang, Y.; Zhang, B.; Tian, Y.; Shao, L. Hrank: Filter pruning using high-rank feature map. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1529–1538. [Google Scholar]
  25. Cai, L.; An, Z.; Yang, C.; Yan, Y.; Xu, Y. Prior gradient mask guided pruning-aware fine-tuning. In Proceedings of the AAAI Conference on Artificial Intelligence, virtually, 22 February–1 March 2022; Volume 1. [Google Scholar]
  26. Huang, Z.; Shao, W.; Wang, X.; Lin, L.; Luo, P. Convolution-weight distribution assumption: Rethinking the criteria of channel pruning. arXiv 2020, arXiv:2004.11627. [Google Scholar]
  27. Mozer, M.C.; Smolensky, P. Skeletonization: A technique for trimming the fat from a network via relevance assessment. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA; 1988; Volume 1. [Google Scholar]
  28. Han, S.; Pool, J.; Tran, J.; Dally, W. Learning both weights and connections for efficient neural network. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  29. Han, S.; Mao, H.; Dally, W.J. Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding. arXiv 2015, arXiv:1510.00149. [Google Scholar]
  30. Wen, W.; Wu, C.; Wang, Y.; Chen, Y.; Li, H. Learning structured sparsity in deep neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  31. Liu, B.; Wang, M.; Foroosh, H.; Tappen, M.; Pensky, M. Sparse convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 806–814. [Google Scholar]
  32. Ye, J.; Lu, X.; Lin, Z.; Wang, J.Z. Rethinking the smaller-normless-informative assumption in channel pruning of convolution layers. arXiv 2018, arXiv:1802.00124. [Google Scholar]
  33. Lee, N.; Ajanthan, T.; Torr, P.H. Snip: Single-shot network pruning based on connection sensitivity. arXiv 2018, arXiv:1810.02340. [Google Scholar]
  34. Luo, J.-H.; Wu, J.; Lin, W. Thinet: A filter level pruning method for deep neural network compression. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5058–5066. [Google Scholar]
  35. He, Y.; Zhang, X.; Sun, J. Channel pruning for accelerating very deep neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1389–1397. [Google Scholar]
  36. Parkash, V.; Carcangiu, M.L. Endometrioid endometrial adenocarcinoma with psammoma bodies. Am. J. Surg. Pathol. 1997, 21, 399–406. [Google Scholar] [CrossRef]
  37. Ashok, A.; Rhinehart, N.; Beainy, F.; Kitani, K.M. N2n learning: Network to network compression via policy gradient reinforcement learning. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  38. Peng, H.; Wu, J.; Chen, S.; Huang, J. Collaborative channel pruning for deep networks. In Proceedings of the International Conference on Machine Learning (PMLR), Long Beach, CA, USA, 9–15 June 2019; pp. 5113–5122. [Google Scholar]
  39. Chen, W.; Wilson, J.; Tyree, S.; Weinberger, K.; Chen, Y. Compressing neural networks with the hashing trick. In Proceedings of the International Conference on Machine Learning (PMLR), Lille, France, 7–9 July 2015; pp. 2285–2294. [Google Scholar]
  40. Xu, Y.; Wang, Y.; Zhou, A.; Lin, W.; Xiong, H. Deep neural network compression with single and multiple level quantization. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. No. 1. [Google Scholar]
  41. Rastegari, M.; Ordonez, V.; Redmon, J.; Farhadi, A. Xnor-net: Imagenet classification using binary convolutional neural networks. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 525–542. [Google Scholar]
  42. Courbariaux, M.; Hubara, I.; Soudry, D.; El-Yaniv, R.; Bengio, Y. Binarized neural networks: Training deep neural networks with weights and activations constrained to+ 1 or-1. arXiv 2016, arXiv:1602.02830. [Google Scholar]
  43. Miyashita, D.; Lee, E.H.; Murmann, B. Convolutional neural networks using logarithmic data representation. arXiv 2016, arXiv:1603.01025. [Google Scholar]
  44. Gong, J.; Shen, H.; Zhang, G.; Liu, X.; Li, S.; Jin, G.; Maheshwari, N.; Fomenko, E.; Segal, E. Highly efficient 8-bit low precision inference of convolutional neural networks with intelcaffe. In Proceedings of the 1st on Reproducible Quality-Efficient Systems Tournament on Codesigning Pareto-Efficient Deep Learning, Williamsburg, VA, USA, 24 April 2018; p. 1. [Google Scholar]
  45. Zhao, R.; Hu, Y.; Dotzel, J.; de Sa, C.; Zhang, Z. Improving neural network quantization without retraining using outlier channel splitting. In Proceedings of the International Conference on Machine Learning (PMLR), Long Beach, CA, USA, 9–15 June 2019; pp. 7543–7552. [Google Scholar]
  46. Nagel, M.; Baalen, M.v.; Blankevoort, T.; Welling, M. Datafree quantization through weight equalization and bias correction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, South Korea, 27 October–2 November 2019; pp. 1325–1334. [Google Scholar]
  47. Maddison, C.J.; Mnih, A.; Teh, Y.W. The concrete distribution: A continuous relaxation of discrete random variables. arXiv 2016, arXiv:1611.00712. [Google Scholar]
  48. Hwang, K.; Sung, W. Fixed-point feedforward deep neural network design using weights+ 1, 0, and- 1. In Proceedings of the 2014 IEEE Workshop on Signal Processing Systems (SiPS), Belfast, Ireland, 20–22 October 2014; pp. 1–6. [Google Scholar]
  49. Zhu, X.; Zhou, W.; Li, H. Adaptive layerwise quantization for deep neural network compression. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018; pp. 1–6. [Google Scholar]
  50. Ba, J.; Caruana, R. Do deep nets really need to be deep? In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  51. Li, J.; Zhao, R.; Huang, J.-T.; Gong, Y. Learning small-size dnn with output-distribution-based criteria. In Proceedings of the Fifteenth Annual Conference of the International Speech Communication Association, Singapore, 14–18 September 2014. [Google Scholar]
  52. Ye, J.; Ji, Y.; Wang, X.; Ou, K.; Tao, D.; Song, M. Student becoming the master: Knowledge amalgamation for joint scene parsing, depth estimation, and more. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2829–2838. [Google Scholar]
  53. Yim, J.; Joo, D.; Bae, J.; Kim, J. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4133–4141. [Google Scholar]
  54. Jaderberg, M.; Vedaldi, A.; Zisserman, A. Speeding up convolutional neural networks with low rank expansions. arXiv 2014, arXiv:1405.3866. [Google Scholar]
  55. Wang, P.; Cheng, J. Fixed-point factorized networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4012–4020. [Google Scholar]
  56. Kim, Y.-D.; Park, E.; Yoo, S.; Choi, T.; Yang, L.; Shin, D. Compression of deep convolutional neural networks for fast and low power mobile applications. arXiv 2015, arXiv:1511.06530. [Google Scholar]
  57. Lebedev, V.; Ganin, Y.; Rakhuba, M.; Oseledets, I.; Lempitsky, V. Speeding-up convolutional neural networks using fine-tuned cp-decomposition. arXiv 2014, arXiv:1412.6553. [Google Scholar]
  58. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. 2009. Available online: http://www.cs.utoronto.ca/~kriz/learning-features-2009-TR.pdf (accessed on 8 April 2009).
  59. He, Y.; Kang, G.; Dong, X.; Fu, Y.; Yang, Y. Soft filter pruning for accelerating deep convolutional neural networks. In Proceedings of the IJCAI International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  60. Wang, Z.; Li, C.; Wang, X. Convolutional neural network pruning with structural redundancy reduction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14913–14922. [Google Scholar]
  61. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the International Conference on Machine Learning, (PMLR), Atlanta, GA, USA, 17–19 June 2013; pp. 1139–1147. [Google Scholar]
  62. Dong, X.; Huang, J.; Yang, Y.; Yan, S. More is less: A more complicated network with less inference complexity. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5840–5848. [Google Scholar]
Figure 1. Our proposed pruning framework. First, we traverse layer-by-layer, and use the convolution layer as the unit to cluster the filters in the convolution layer, respectively. After clustering, each filter gets its label and grouping, and redundant filters are removed in the similarity group according to the proposed redundancy criterion.
Figure 1. Our proposed pruning framework. First, we traverse layer-by-layer, and use the convolution layer as the unit to cluster the filters in the convolution layer, respectively. After clustering, each filter gets its label and grouping, and redundant filters are removed in the similarity group according to the proposed redundancy criterion.
Entropy 25 00009 g001
Figure 2. Visualization of filters and feature maps of the first convolutional layer of VGG16. As can be seen from the figure, a convolutional layer has multiple filters with similar expressive abilities, and the feature maps obtained by similar filters after convolution are also similar.
Figure 2. Visualization of filters and feature maps of the first convolutional layer of VGG16. As can be seen from the figure, a convolutional layer has multiple filters with similar expressive abilities, and the feature maps obtained by similar filters after convolution are also similar.
Entropy 25 00009 g002
Figure 3. Pruning of different structures: (a) is a plain structure, where the number of pruned channels between the two layers is not constrained; (b) is a residual block, where each block has the same number of input and output channels, and the number of input and output channels of the middle layer and the upper and lower layers both satisfy specific constraints.
Figure 3. Pruning of different structures: (a) is a plain structure, where the number of pruned channels between the two layers is not constrained; (b) is a residual block, where each block has the same number of input and output channels, and the number of input and output channels of the middle layer and the upper and lower layers both satisfy specific constraints.
Entropy 25 00009 g003
Figure 4. Our method uses VGG16 on the CIFAR10/100 dataset to obtain different pruned structures according to different redundancy.
Figure 4. Our method uses VGG16 on the CIFAR10/100 dataset to obtain different pruned structures according to different redundancy.
Entropy 25 00009 g004
Table 1. Comparison of Pruned VGG16 on CIFAR10/100 Datasets.
Table 1. Comparison of Pruned VGG16 on CIFAR10/100 Datasets.
Model/DataMethodBaseline Top-1 Acc (%)Pruned Top-1 Acc (%)Top-1(↓) Acc (%)FLOPs (↓) (%)Params (↓) (%)
VGG16/CIFAR10L193.5893.310.2734.2064.00
FPGM93.5893.230.3434.2064.00
Ours93.9293.700.2240.9842.46
Taylor93.9291.242.7878.0384.56
Hrank93.9691.232.7376.5092.00
Ours93.9292.491.4387.4991.20
VGG16/CIFAR100L173.4571.212.2450.4450.23
Taylor73.4570.342.3651.4859.89
FPGM73.4571.392.06-48.93
Ours73.4571.911.5454.1162.49
Table 2. Comparison of Pruned ResNet on CIFAR10/CIFAR100 Datasets.
Table 2. Comparison of Pruned ResNet on CIFAR10/CIFAR100 Datasets.
Model/DataMethodBaseline Top-1 Acc (%)Pruned Top-1 Acc (%)Top-1 (↓) Acc (%)FLOPs (↓) (%)
ResNet-32/CIFAR10L191.8280.0111.8143.76
SFP91.3391.60+0.2753.16
FPGM91.3391.90+0.5753.16
Ours91.8292.11+0.2955.36
ResNet-56/CIFAR10L193.0491.311.7527.60
SFP93.5992.261.3352.60
FPGM93.5992.890.7052.60
HRank93.2693.170.0950.00
SRR-GR93.3893.75+0.3753.80
Ours92.5593.09+0.5460.10
ResNet-110/CIFAR10L193.5392.940.6138.60
SFP93.6893.380.3040.80
FPGM93.6893.73+0.0552.30
Hrank93.5092.650.8568.60
Ours93.6093.170.4370.59
ResNet-32/CIFAR100L166.4858.118.3743.76
SFP66.4864.272.2153.16
FPGM66.4866.640.1653.16
Ours66.4866.87+0.3950.51
ResNet-56/CIFAR100SFP69.0868.031.0563.16
FPGM69.0867.751.3363.16
PGMPF72.9270.212.7152.6
Ours69.0868.570.5163.48
ResNet-110/CIFAR100Ours71.2670.280.9857.73
Table 3. Comparison of Pruned ResNet on ImageNet.
Table 3. Comparison of Pruned ResNet on ImageNet.
Model/DataMethodBaseline Top-1 Acc (%)Pruned Top-1 Acc (%)Top-1 (↓) Acc (%)Baseline Top-5 Acc (%)Pruned Top-5 Acc (%)Top-5 (↓) Acc (%)FLOPs (↓) (%)
ResNet-18MIL69.9866.333.6586.9489.242.3034.6
SFP70.2867.103.1889.6387.781.8541.8
FPGM70.2867.812.4789.6388.111.5241.8
PGMPF70.2366.673.5689.5187.362.1553.5
Ours70.4868.661.8289.6088.441.1653.8
ResNet-34MIL73.4272.990.4391.3691.190.1724.8
L173.2372.171.06---24.2
SFP73.9271.832.0991.6290.331.2941.1
FPGM73.9272.111.8191.6290.690.9341.1
PGMPF73.2770.642.6391.4389.871.5652.7
Ours73.9072.551.3591.5990.790.8052.1
ResNet-50ThiNet75.3074.031.2792.2092.110.0936.79
SFP76.1574.611.5492.8792.060.8141.8
FPGM76.1575.031.1292.8792.400.4742.2
HRank76.1574.981.1792.8792.330.5443.76
SRR-GR76.1375.760.3792.8692.600.1944.10
PGMPF76.0175.110.9092.9392.410.5253.5
Ours75.8272.470.3592.9592.680.2753.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, T.; Song, C.; Zeng, P.; Xia, C. Cluster-Based Structural Redundancy Identification for Neural Network Compression. Entropy 2023, 25, 9. https://doi.org/10.3390/e25010009

AMA Style

Wu T, Song C, Zeng P, Xia C. Cluster-Based Structural Redundancy Identification for Neural Network Compression. Entropy. 2023; 25(1):9. https://doi.org/10.3390/e25010009

Chicago/Turabian Style

Wu, Tingting, Chunhe Song, Peng Zeng, and Changqing Xia. 2023. "Cluster-Based Structural Redundancy Identification for Neural Network Compression" Entropy 25, no. 1: 9. https://doi.org/10.3390/e25010009

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop