A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2019; you can also visit the original URL.
The file type is application/pdf
.
Cold Fusion: Training Seq2Seq Models Together with Language Models
2018
Interspeech 2018
Sequence-to-sequence (Seq2Seq) models with attention have excelled at tasks which involve generating natural language sentences such as machine translation, image captioning and speech recognition. Performance has further been improved by leveraging unlabeled data, often in the form of a language model. In this work, we present the Cold Fusion method, which leverages a pre-trained language model during training, and show its effectiveness on the speech recognition task. We show that Seq2Seq
doi:10.21437/interspeech.2018-1392
dblp:conf/interspeech/SriramJSC18
fatcat:ts3vdbyvcvdtrpihjwdk6g7mfy