CASIA OpenIR  > 毕业生  > 硕士学位论文
高效卷积神经网络设计研究
莫子韬
2020-05-25
页数78
学位类型硕士
中文摘要

近年来,深度学习在计算机视觉、自然语言处理、语音识别等领域取得了巨大的成功,其中一个原因要归功于强大的深度神经网络。深度卷积神经网络是其中一类重要代表,已经成为现代计算机视觉里最常用的模型。然而,深度卷积神经网络的参数量和计算量巨大,往往无法满足实际应用的需求,设计更加高效的神经网络已成为一个重要的研究课题。本文分别针对特定任务卷积神经网络的设计、量化卷积神经网络的设计等问题,从手工设计、自动化搜索两个角度出发,探索更加高效的网络结构。本文的主要贡献总结如下:

针对单图像超分辨率这个任务,提出了一种基于数值微分方程的启发式设计模式,通过建立数值微分方程与卷积神经网络之间的联系,将 Leapfrog 方法、二阶龙格库塔法和三阶龙格库塔法对应的数值迭代公式直接映射为新的网络模块。通过这种方式得到的网络 OISR-RK3 在相近的计算量下超越了之前性能最好的工作。

针对单图像超分辨率网络模型计算量大的问题,引入轻量化深度可分离卷积以及金字塔结构,根据单图像超分辨率任务的特点提出了设计金字塔结构的参考准则,并且相应地提出了一种梯形金字塔网络,实验结果表明它能够在效率和性能之间取得更好的平衡。

针对卷积神经网络二值化性能损失过大的问题,引入基于权值共享的高效的神经网络架构搜索的方法,并且设计了一个适用于二值化卷积神经网络的搜索空间,解决了超网训练不稳定的问题。使用二值化神经网络架构搜索得到的网络与之前的研究相比,能够在计算量相近的情况下取得最佳性能,在 ImageNet 上达到 60.2% 的精度。

英文摘要

In recent years, deep learning has achieved great success in computer vision, natural language processing and speech recognition with the help of powerful deep neural networks. As one of the most important representatives, deep convolutional neural networks have become the most commonly used model in modern computer vision. However, the huge amount of parameters and computation impedes their wider application, designing more efficient neural networks has become an important topic. This thesis focuses on task-specific convolutional neural networks and quantized convolutional neural networks. We explore efficient structures from the perspective of manual design and neural architecture search, respectively. The main contributions of this thesis are summarized as follows:


In order to obtain better networks for single image super-resolution without overly relying on trial and error, we introduce ODE-Inspired design scheme to this task. By establishing the relationship between numerical differential equations and convolutional neural networks, Leapfrog method, second-order Runge-Kuta and third-order Runge-Kutta method are directly mapped to the new building blocks of convolutional neural networks. The networks obtained in this way can outperform prior arts with similar computation overhead. 


In order to further reduce computation of the networks for single image super-resolution, we introduce depthwise separable convolution and feature pyramid structure. According to the characteristics of this low-level vision task, several criteria for designing pyramid structure are proposed, with which we develop ladder pyramid networks. Experiments demonstrate that the proposed ladder pyramid networks can achieve a better balance between efficiency and performance.


In order to automate the design flow and explore better architectures for binarized neural networks, we introduce efficient one-shot method of neural architecture search. We develop a search space dedicated for binarized convolutional neural networks, which can stabilize the training of the supernet. The searched binarized neural network can outperform previous work while keeping similar amount of computation, reaching 60.2% accuracy on ImageNet dataset.
 

关键词高效卷积神经网络,单图像超分辨率,特征金字塔网络,二值化神经网络,神经网络架构搜索
学科门类工学
语种中文
七大方向——子方向分类AI芯片与智能计算
文献类型学位论文
条目标识符http://ir.ia.ac.cn/handle/173211/40123
专题毕业生_硕士学位论文
作者单位中国科学院自动化研究所
第一作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
莫子韬. 高效卷积神经网络设计研究[D]. 中国科学院自动化研究所. 中国科学院大学,2020.
条目包含的文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
Thesis.pdf(2579KB)其他 限制开放CC BY-NC-SA
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[莫子韬]的文章
百度学术
百度学术中相似的文章
[莫子韬]的文章
必应学术
必应学术中相似的文章
[莫子韬]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。