Towards Accurate Post-training Network Quantization via Bit-Split and Stitching
Wang, Peisong1,2; Chen, Qiang1,2; He, Xiangyu1,2; Cheng, Jian1,2
2020
会议名称International Conference on Machine Learning
期号119
页码9847-9856
会议日期2020
会议地点Online
摘要

Network quantization is essential for deploying deep models to IoT devices due to its high efficiency. Most existing quantization approaches rely on the full training datasets and the timeconsuming fine-tuning to retain accuracy. Posttraining quantization does not have these problems, however, it has mainly been shown effective for 8-bit quantization due to the simple optimization strategy. In this paper, we propose a Bit-Split and Stitching framework (Bit-split) for lower-bit post-training quantization with minimal accuracy degradation. The proposed framework is validated on a variety of computer vision tasks, including image classification, object detection, instance segmentation, with various network architectures. Specifically, Bit-split can achieve near-original model performance even when quantizing FP32 models to INT3 without fine-tuning.

收录类别EI
七大方向——子方向分类AI芯片与智能计算
文献类型会议论文
条目标识符http://ir.ia.ac.cn/handle/173211/40620
专题紫东太初大模型研究中心_图像与视频分析
复杂系统认知与决策实验室_高效智能计算与学习
通讯作者Cheng, Jian
作者单位1.Institute of Automation, Chinese Academy of Sciences
2.University of Chinese Academy of Sciences
第一作者单位中国科学院自动化研究所
通讯作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
Wang, Peisong,Chen, Qiang,He, Xiangyu,et al. Towards Accurate Post-training Network Quantization via Bit-Split and Stitching[C],2020:9847-9856.
条目包含的文件 下载所有文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
Towards Accurate Pos(389KB)会议论文 开放获取CC BY-NC-SA浏览 下载
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Wang, Peisong]的文章
[Chen, Qiang]的文章
[He, Xiangyu]的文章
百度学术
百度学术中相似的文章
[Wang, Peisong]的文章
[Chen, Qiang]的文章
[He, Xiangyu]的文章
必应学术
必应学术中相似的文章
[Wang, Peisong]的文章
[Chen, Qiang]的文章
[He, Xiangyu]的文章
相关权益政策
暂无数据
收藏/分享
文件名: Towards Accurate Post-training Network Quantization via Bit-Split and Stitching.pdf
格式: Adobe PDF
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。