Title
Pre-training on high-resource speech recognition improves low-resource speech-to-text translation.
Abstract
We present a simple approach to improve direct speech-to-text translation (ST) when the source language is low-resource: we pre-train the model on a high-resource automatic speech recognition (ASR) task, and then fine-tune its parameters for ST. We demonstrate that our approach is effective by pre-training on 300 hours of English ASR data to improve Spanish-English ST from 10.8 to 20.2 BLEU when only 20 hours of Spanish-English ST training data is available. Through an ablation study, we find that the pre-trained encoder (acoustic model) accounts for most of the improvement, which is surprising since the shared language in these tasks is the target language (text), and not the source language (audio). Applying this insight, we show that pre-training on ASR helps ST even when the ASR language differs from both source and target ST languages: pre-training on French ASR also improves Spanish-English ST. Finally, we show that the approach improves a true low-resource task: pre-training on a combination of English ASR and French ASR improves Mboshi-French ST, where only 4 hours of data are available, from 3.5 to 7.1 BLEU.
Year
DOI
Venue
2018
10.18653/v1/n19-1006
north american chapter of the association for computational linguistics
Field
DocType
Volume
Training set,Computer science,Speech recognition,Natural language processing,Encoder,Artificial intelligence,Acoustic model
Journal
abs/1809.01431
Citations 
PageRank 
References 
1
0.36
23
Authors
5
Name
Order
Citations
PageRank
Sameer Bansal131.07
Herman Kamper215020.70
Karen Livescu3125471.43
Adam Lopez453834.69
Sharon Goldwater51437103.96