End-to-End Speech Translation with Knowledge Distillation
Yuchen Liu3,4; Hao Xiong2; Jiajun Zhang3,4; Zhongjun He2; Hua Wu2; Haifeng Wang2; Chengqing Zong1,3,4
2019-09
会议日期Sep. 15-19, 2019
会议地点Graz,Austria
英文摘要

End-to-end speech translation (ST), which directly translates from source language speech into target language text, has attracted intensive attentions in recent years. Compared to conventional pipeline systems, end-to-end ST model has potential benefits of lower latency, smaller model size and less error propagation. However, it is notoriously difficult to implement such model which combines automatic speech recognition (ASR) and machine translation (MT) together. In this paper, we propose a knowledge distillation approach to improve ST by transferring the knowledge from text translation. Specifically, we first train a text translation model, regarded as the teacher model, and then ST model is trained to learn the output probabilities of teacher model through knowledge distillation. Experiments on English-French Augmented LibriSpeech and English-Chinese TED corpus show that end-to-end ST is possible to implement on both similar and dissimilar language pairs. In addition, with the instruction of the teacher model, end-to end ST model can gain significant improvements by over 3.5 BLEU points. 

内容类型会议论文
源URL[http://ir.ia.ac.cn/handle/173211/44410]  
专题模式识别国家重点实验室_自然语言处理
作者单位1.CAS Center for Excellence in Brain Science and Intelligence Technology
2.Baidu Inc.
3.NLPR, Institute of Automation, Chinese Academy of Sciences
4.University of Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Yuchen Liu,Hao Xiong,Jiajun Zhang,et al. End-to-End Speech Translation with Knowledge Distillation[C]. 见:. Graz,Austria. Sep. 15-19, 2019.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace