Subject: Computer Science - Computation and Language | Statistics - Machine Learning | Computer Science - Machine Learning
Sequence-to-sequence attention-based models on subword units allow simple open-vocabulary end-to-end speech recognition. In this work, we show that such models can achieve competitive results on the Switchboard 300h and LibriSpeech 1000h tasks. In particular, we report ... View more
 A. Zeyer, P. Doetsch, P. Voigtlaender, R. Schlu¨ter, and H. Ney, “A comprehensive study of deep bidirectional LSTM RNNs for acoustic modeling in speech recognition,” in ICASSP, New Orleans, LA, USA, Mar. 2017, pp. 2462-2466.
 H. Bourlard and N. Morgan, Connectionist speech recognition: a hybrid approach. Springer, 1994, vol. 247.
 A. J. Robinson, “An application of recurrent nets to phone probability estimation,” Neural Networks, IEEE Transactions on, vol. 5, no. 2, pp. 298-305, 1994.
 S. J. Young, J. J. Odell, and P. C. Woodland, “Tree-based state tying for high accuracy acoustic modelling,” in Proceedings of the workshop on Human Language Technology. Association for Computational Linguistics, 1994, pp. 307-312.
 A. Senior, G. Heigold, M. Bacchiani, and H. Liao, “GMM-free DNN acoustic model training,” in ICASSP, 2014.
 A. Zeyer, E. Beck, R. Schlu¨ter, and H. Ney, “CTC in the context of generalized full-sum HMM training,” in Interspeech, Stockholm, Sweden, Aug. 2017, pp. 944-948.
 D. Povey, V. Peddinti, D. Galvez, P. Ghahremani, V. Manohar, X. Na, Y. Wang, and S. Khudanpur, “Purely sequence-trained neural networks for ASR based on lattice-free MMI,” in Interspeech, 2016, pp. 2751-2755.
 H. Sak, A. Senior, K. Rao, and F. Beaufays, “Fast and accurate recurrent neural network acoustic models for speech recognition,” in Interspeech, 2015.
 S. Kanthak and H. Ney, “Context-dependent acoustic modeling using graphemes for large vocabulary speech recognition,” in ICASSP, Orlando, FL, USA, May 2002, pp. 845-848.
 A. Graves, S. Ferna´ndez, F. Gomez, and J. Schmidhuber, “Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks,” in ICML. ACM, 2006, pp. 369-376.