Self-Attention Transducers for End-to-End Speech Recognition
Zhengkun Tian1,2; Jiangyan Yi1; Jianhua Tao1,2,3; Ye Bai1,2; Zhengqi Wen1
2019-09
会议名称INTERSPEECH
会议日期September 15–19, 2019
会议地点Graz, Austria
摘要

Recurrent neural network transducers (RNN-T) have been successfully
applied in end-to-end speech recognition. However, the recurrent structure makes it difficult for parallelization. In this paper, we propose a self-attention transducer (SA-T) for speech recognition. RNNs are replaced with self-attention blocks, which are powerful to model long-term dependencies inside sequences and able to be efficiently parallelized. Furthermore, a path-aware regularization is proposed to assist SA-T to learn alignments and improve the performance. Additionally, a chunk-flow mechanism is utilized to achieve online decoding. All experiments are conducted on a Mandarin Chinese dataset AISHELL-1. The results demonstrate that our proposed approach
achieves a 21.3% relative reduction in character error rate compared with the baseline RNN-T. In addition, the SA-T with chunk-flow mechanism can perform online decoding with only a little degradation of the performance.

收录类别EI
语种英语
文献类型会议论文
条目标识符http://ir.ia.ac.cn/handle/173211/48608
专题多模态人工智能系统全国重点实验室_智能交互
作者单位1.National Laboratory of Pattern Recognition, Institute of Automation, CASIA
2.School of Artificial Intelligence, University of Chinese Academy of Sciences
3.CAS Center for Excellence in Brain Science and Intelligence Technology
第一作者单位模式识别国家重点实验室
推荐引用方式
GB/T 7714
Zhengkun Tian,Jiangyan Yi,Jianhua Tao,et al. Self-Attention Transducers for End-to-End Speech Recognition[C],2019.
条目包含的文件 下载所有文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
tian19b_interspeech.(278KB)会议论文 开放获取CC BY-NC-SA浏览 下载
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Zhengkun Tian]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
百度学术
百度学术中相似的文章
[Zhengkun Tian]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
必应学术
必应学术中相似的文章
[Zhengkun Tian]的文章
[Jiangyan Yi]的文章
[Jianhua Tao]的文章
相关权益政策
暂无数据
收藏/分享
文件名: tian19b_interspeech.pdf
格式: Adobe PDF
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。