Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Ye Bai; Jiangyan Yi; Jianhua Tao; Zhengkun Tian; Zhengqi Wen; Shuai Zhang
2020
会议名称interspeecch
会议日期2020
会议地点shanghai
摘要

Although attention based end-to-end models have achieved
promising performance in speech recognition, the multi-pass
forward computation in beam-search increases inference time
cost, whichlimitstheirpracticalapplications. Toaddressthisis-
sue, we propose a non-autoregressive end-to-end speech recog-
nition system called LASO (listen attentively, and spell once).
Because of the non-autoregressive property, LASO predicts a
textual token in the sequence without the dependence on other
tokens. Without beam-search, the one-pass propagation much
reduces inference time cost of LASO. And because the model
is based on the attention based feedforward structure, the com-
putation can be implemented in parallel efficiently. We conduct
experiments on publicly available Chinese dataset AISHELL-
1. LASO achieves a character error rate of 6.4%, which out-
performs the state-of-the-art autoregressive transformer model
(6.7%). The average inference latency is 21 ms, which is 1/50
of the autoregressive transformer model.

七大方向——子方向分类语音识别与合成
文献类型会议论文
条目标识符http://ir.ia.ac.cn/handle/173211/44978
专题多模态人工智能系统全国重点实验室_智能交互
作者单位Institute of Automation, Chinese Academy of Sciences
第一作者单位中国科学院自动化研究所
推荐引用方式
GB/T 7714
Ye Bai,Jiangyan Yi,Jianhua Tao,et al. Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition[C],2020.
条目包含的文件 下载所有文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
LASO-camera.pdf(801KB)会议论文 开放获取CC BY-NC-SA浏览 下载
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Ye Bai]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
百度学术
百度学术中相似的文章
[Ye Bai]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
必应学术
必应学术中相似的文章
[Ye Bai]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
相关权益政策
暂无数据
收藏/分享
文件名: LASO-camera.pdf
格式: Adobe PDF
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。