CASIA OpenIR  > 模式识别国家重点实验室  > 智能交互
Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT
Ye Bai; Jiangyan Yi; Jianhua Tao; Zhengkun Tian; Zhengqi Wen; Shuai Zhang
Source PublicationIEEE/ACM Transactions on Audio, Speech, and Language Processing
2021
Issue29Pages:1897 - 1911
Abstract

Attention-based encoder-decoder (AED) models
have achieved promising performance in speech recognition.
However, because the decoder predicts text tokens (such as
characters or words) in an autoregressive manner, it is difficult
for an AED model to predict all tokens in parallel. This makes
the inference speed relatively slow. In contrast, we propose an
end-to-end non-autoregressive speech recognition model called
LASO (Listen Attentively, and Spell Once). The model aggre-
gates encoded speech features into the hidden representations
corresponding to each token with attention mechanisms. Thus,
the model can capture the token relations by self-attention on
the aggregated hidden representations from the whole speech
signal rather than autoregressive modeling on tokens. Without
explicitly autoregressive language modeling, this model predicts
all tokens in the sequence in parallel so that the inference is
efficient. Moreover, we propose a cross-modal transfer learning
method to use a text-modal language model to improve the
performance of speech-modal LASO by aligning token semantics.
We conduct experiments on two scales of public Chinese speech
datasets AISHELL-1 and AISHELL-2. Experimental results
show that our proposed model achieves a speedup of about 50×
and competitive performance, compared with the autoregressive
transformer models. And the cross-modal knowledge transferring
from the text-modal model can improve the performance of the
speech-modal model.

Keyword端到端语音识别、迁移学习、知识蒸馏、老师-学生学习、BERT、非自回归语音识别
Sub direction classification语音识别与合成
Document Type期刊论文
Identifierhttp://ir.ia.ac.cn/handle/173211/44977
Collection模式识别国家重点实验室_智能交互
AffiliationInstitute of Automation, Chinese Academy of Sciences
First Author AffilicationInstitute of Automation, Chinese Academy of Sciences
Recommended Citation
GB/T 7714
Ye Bai,Jiangyan Yi,Jianhua Tao,et al. Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT[J]. IEEE/ACM Transactions on Audio, Speech, and Language Processing,2021(29):1897 - 1911.
APA Ye Bai,Jiangyan Yi,Jianhua Tao,Zhengkun Tian,Zhengqi Wen,&Shuai Zhang.(2021).Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT.IEEE/ACM Transactions on Audio, Speech, and Language Processing(29),1897 - 1911.
MLA Ye Bai,et al."Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT".IEEE/ACM Transactions on Audio, Speech, and Language Processing .29(2021):1897 - 1911.
Files in This Item:
File Name/Size DocType Version Access License
turn2-v0.2.pdf(1163KB)期刊论文作者接受稿开放获取CC BY-NC-SAView
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Ye Bai]'s Articles
[Jiangyan Yi]'s Articles
[Jianhua Tao]'s Articles
Baidu academic
Similar articles in Baidu academic
[Ye Bai]'s Articles
[Jiangyan Yi]'s Articles
[Jianhua Tao]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Ye Bai]'s Articles
[Jiangyan Yi]'s Articles
[Jianhua Tao]'s Articles
Terms of Use
No data!
Social Bookmark/Share
File name: turn2-v0.2.pdf
Format: Adobe PDF
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.