A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
ASAPP-ASR: Multistream CNN and Self-Attentive SRU for SOTA Speech Recognition
2020
Interspeech 2020
In this paper we present state-of-the-art (SOTA) performance on the LibriSpeech corpus with two novel neural network architectures, a multistream CNN for acoustic modeling and a selfattentive simple recurrent unit (SRU) for language modeling. In the hybrid ASR framework, the multistream CNN acoustic model processes an input of speech frames in multiple parallel pipelines where each stream has a unique dilation rate for diversity. Trained with the SpecAugment data augmentation method, it
doi:10.21437/interspeech.2020-2947
dblp:conf/interspeech/PanSWH0M20
fatcat:pfvnsjbilfdzhmadi24doeql5e