Subject: Computer Science - Computation and Language | Computer Science - Sound | Electrical Engineering and Systems Science - Audio and Speech Processing | Computer Science - Machine Learning
Grapheme-to-phoneme (G2P) conversion is an important task in automatic speech recognition and text-to-speech systems. Recently, G2P conversion is viewed as a sequence to sequence task and modeled by RNN or CNN based encoder-decoder framework. However, previous works do ... View more
 M. Bisani and H. Ney, “Joint-sequence models for grapheme-tophoneme conversion,” Speech communication, vol. 50, no. 5, pp. 434-451, 2008.
 M.-j. Chae, K. Park, L. Bang, S. Suh, L. Park, N. Kimt, and L. Park, “Convolutional sequence to sequence model with nonsequential greedy decoding for grapheme to phoneme conversion,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 2486- 2490.
 G. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,” arXiv preprint arXiv:1503.02531, 2015.
 T. Furlanello, Z. C. Lipton, M. Tschannen, L. Itti, and A. Anandkumar, “Born again neural networks,” arXiv preprint arXiv:1805.04770, 2018.
 Y. Kim and A. M. Rush, “Sequence-level knowledge distillation,” arXiv preprint arXiv:1606.07947, 2016.
 M. Freitag, Y. Al-Onaizan, and B. Sankaran, “Ensemble distillation for neural machine translation,” arXiv preprint arXiv:1702.01802, 2017.
 X. Tan, Y. Ren, D. He, T. Qin, and T.-Y. Liu, “Multilingual neural machine translation with knowledge distillation,” in International Conference on Learning Representations, 2019. [Online]. Available: https://openreview.net/forum?id=S1gUsoR9YX
 A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems, 2017, pp. 5998-6008.
 C. Gong, X. Tan, D. He, and T. Qin, “Sentence-wise smooth regularization for sequence to sequence learning,” 2018.
 S. Zhou, L. Dong, S. Xu, and B. Xu, “A comparison of modeling units in sequence-to-sequence speech recognition with the transformer on mandarin chinese,” in International Conference on Neural Information Processing. Springer, 2018, pp. 210-220.