Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Zhang, Wei | Wang, Zhiming; *
Affiliations: School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, P.R. China
Correspondence: [*] Corresponding author. School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, 100083, P.R. China. E-mail: wangzhiming@ustb.edu.cn.
Abstract: Deep Convolutional Neural Networks (CNNs) have been widely used in various domains due to their outstanding performance. However, they simultaneously bring enormous computational overhead, making it difficult to deploy to mobile and edge devices. Therefore, researchers use network compression techniques such as quantization, knowledge distillation and neural network pruning to alleviate this problem. Among network pruning, filter pruning has received broad attention. At present, most of the filter pruning methods need to define pruning rates manually, which is a trial-and-error process and requires rich experimental experience. Some methods obtain global optimal network parameters by Neural Architecture Search (NAS) or Evolutionary Algorithms (EA) to overcome this difficulty. However, they also introduce huge computational burden. To mitigate the above problems, this study proposes a pruning strategy based on Principal Component Analysis (PCA) called PCA-Pruner. Filter weights of a layer is regarded as a set of features, and the number of filters responding to feature dimension. Then, the number of reserved filters in each layer can be determined by PCA which is a classical dimensionality reduction technology. After that, we calculate the L1 norm of each filter in each layer and use it as an importance measurement to prune filters. Experimental results show that PCA-Pruner achieve performance improvements over the state-of-the-arts algorithms. For example, we compress the FLOPs and parameters of ResNet-56 on CIFA-10 by 45.8% and 47.1%, with an increase in accuracy of 0.27%. For ResNet-110 on CIFAR-10, we improve the accuracy by 0.58% and reduce the FLOPs and Params of the model by 58.3% and 56.2%, respectively. Towards ResNet-56 on CIFAR-100 dataset, we achieve a 38.8% FLOPs decrease and 38.0% Params reduction with only 0.69% accuracy loss.
Keywords: Network compression, neural network pruning, dimensionality reduction, PCA
DOI: 10.3233/JIFS-211555
Journal: Journal of Intelligent & Fuzzy Systems, vol. 43, no. 4, pp. 4803-4813, 2022
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
sales@iospress.com
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
info@iospress.nl
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office info@iospress.nl
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
china@iospress.cn
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
如果您在出版方面需要帮助或有任何建, 件至: editorial@iospress.nl