DPFPS: Dynamic and Progressive Filter Pruning for Compressing Convolutional Neural Networks from Scratch
Ruan, Xiaofeng1,2; Liu, Yufan1,2; Li, Bing1,4; Yuan, Chunfeng1; Hu, Weiming1,2,3
2021-05-18
会议名称The Thirty-Fifth AAAI Conference on Artificial Intelligence
会议日期2021.2.2-2021.2.9
会议地点virtual conference
摘要

Filter pruning is a commonly used method for compressing Convolutional Neural Networks (ConvNets), due to its friendly hardware supporting and flexibility. However, existing methods mostly need a cumbersome procedure, which brings many extra hyper-parameters and training epochs. This is because only using sparsity and pruning stages cannot obtain a satisfying performance. Besides, many works do not consider the difference of pruning ratio across different layers. To overcome these limitations, we propose a novel dynamic and progressive filter pruning (DPFPS) scheme that directly learns a structured sparsity network from Scratch. In particular, DPFPS imposes a new structured sparsityinducing regularization specifically upon the expected pruning parameters in a dynamic sparsity manner. The dynamic sparsity scheme determines sparsity allocation ratios of different layers and a Taylor series based channel sensitivity criteria is presented to identify the expected pruning parameters. Moreover, we increase the structured sparsity-inducing penalty in a progressive manner. This helps the model to be sparse gradually instead of forcing the model to be sparse at the beginning. Our method solves the pruning ratio based optimization problem by an iterative soft-thresholding algorithm (ISTA) with dynamic sparsity. At the end of the training, we only need to remove the redundant parameters without other stages, such as fine-tuning. Extensive experimental results show that the proposed method is competitive with 11 state-of-the-art methods on both small-scale and large-scale datasets (i.e., CIFAR and ImageNet). Specifically, on ImageNet, we achieve a 44.97% pruning ratio of FLOPs by compressing ResNet-101, even with an increase of 0.12% Top-5 accuracy. Our pruned models and codes are released at https://github.com/taoxvzi/DPFPS.

收录类别EI
语种英语
七大方向——子方向分类机器学习
文献类型会议论文
条目标识符http://ir.ia.ac.cn/handle/173211/44803
专题多模态人工智能系统全国重点实验室_视频内容安全
通讯作者Li, Bing
作者单位1.National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences
2.School of Artificial Intelligence, University of Chinese Academy of Sciences
3.CAS Center for Excellence in Brain Science and Intelligence Technology
4.PeopleAI Inc.
第一作者单位模式识别国家重点实验室
通讯作者单位模式识别国家重点实验室
推荐引用方式
GB/T 7714
Ruan, Xiaofeng,Liu, Yufan,Li, Bing,et al. DPFPS: Dynamic and Progressive Filter Pruning for Compressing Convolutional Neural Networks from Scratch[C],2021.
条目包含的文件 下载所有文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
DPFPS_Dynamic and Pr(652KB)会议论文 开放获取CC BY-NC-SA浏览 下载
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Ruan, Xiaofeng]的文章
[Liu, Yufan]的文章
[Li, Bing]的文章
百度学术
百度学术中相似的文章
[Ruan, Xiaofeng]的文章
[Liu, Yufan]的文章
[Li, Bing]的文章
必应学术
必应学术中相似的文章
[Ruan, Xiaofeng]的文章
[Liu, Yufan]的文章
[Li, Bing]的文章
相关权益政策
暂无数据
收藏/分享
文件名: DPFPS_Dynamic and Progressive Filter Pruning for Compressing Convolutional Neural Networks from Scratch.pdf
格式: Adobe PDF
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。