papers AI Learner
The Github is limit! Click to go to the new site.

Improving the Robustness of Speech Translation

2018-11-02
Xiang Li, Haiyang Xue, Wei Chen, Yang Liu, Yang Feng, Qun Liu

Abstract

Although neural machine translation (NMT) has achieved impressive progress recently, it is usually trained on the clean parallel data set and hence cannot work well when the input sentence is the production of the automatic speech recognition (ASR) system due to the enormous errors in the source. To solve this problem, we propose a simple but effective method to improve the robustness of NMT in the case of speech translation. We simulate the noise existing in the realistic output of the ASR system and inject them into the clean parallel data so that NMT can work under similar word distributions during training and testing. Besides, we also incorporate the Chinese Pinyin feature which is easy to get in speech translation to further improve the translation performance. Experiment results show that our method has a more stable performance and outperforms the baseline by an average of 3.12 BLEU on multiple noisy test sets, even while achieves a generalization improvement on the WMT’17 Chinese-English test set.

Abstract (translated by Google)
URL

https://arxiv.org/abs/1811.00728

PDF

https://arxiv.org/pdf/1811.00728


Similar Posts

Comments