R. J. Weiss, J. Chorowski, N. Jaitly, Y. Wu, and Z. Chen, “Sequence-to-sequence models can directly translate foreign speech,” Proc. Interspeech 2017, pp. 2625-2629, 2017. [OpenAIRE]
 A. Bérard, O. Pietquin, L. Besacier, and C. Servan, “Listen and translate: A proof of concept for end-toend speech-to-text translation,” in NIPS Workshop on end-to-end learning for speech and audio processing, 2016. [OpenAIRE]
 S. Peitz, S. Wiesler, M. Nußbaum-Thom, and H. Ney, “Spoken language translation using automatically transcribed text in training,” in International Workshop on Spoken Language Translation (IWSLT) 2012, 2012. [OpenAIRE]
 M. Sperber, G. Neubig, J. Niehues, and A. Waibel, “Neural lattice-to-sequence models for uncertain inputs,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017, pp. 1380-1389. [OpenAIRE]
 M. Post, G. Kumar, A. Lopez, D. Karakos, C. CallisonBurch, and S. Khudanpur, “Improved speech-to-text translation with the fisher and callhome spanishenglish speech translation corpus,” in Proc. IWSLT, 2013.
 N. Jan, R. Cattoni, S. Sebastian, M. Cettolo, M. Turchi, and M. Federico, “The iwslt 2018 evaluation campaign,” in International Workshop on Spoken Language Translation, 2018, pp. 2-6.
 X. Li, H. Xue, W. Chen, Y. Liu, Y. Feng, and Q. Liu, “Improving the robustness of speech translation,” arXiv preprint arXiv:1811.00728, 2018.
 M. Sperber, J. Niehues, and A. Waibel, “Toward robust neural machine translation for noisy input sequences,” in International Workshop on Spoken Language Translation (IWSLT), Tokyo, Japan, 2017. [OpenAIRE]
 A. Bérard, L. Besacier, A. C. Kocabiyikoglu, and O. Pietquin, “End-to-end automatic speech translation of audiobooks,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 6224-6228. [OpenAIRE]
 D. Serdyuk, Y. Wang, C. Fuegen, A. Kumar, B. Liu, and Y. Bengio, “Towards end-to-end spoken language understanding,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 5754-5758. [OpenAIRE]
 L. Duong, A. Anastasopoulos, D. Chiang, S. Bird, and T. Cohn, “An attentional model for speech translation without transcription,” in Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2016, pp. 949-959. [OpenAIRE]
 S. Bansal, H. Kamper, A. Lopez, and S. Goldwater, “Towards speech-to-text translation without speech recognition,” EACL 2017, p. 474, 2017.
 Y. Belinkov and Y. Bisk, “Synthetic and natural noise both break neural machine translation,” arXiv preprint arXiv:1711.02173, 2017. [OpenAIRE]
 N.-T. Le, B. Lecouteux, and L. Besacier, “Disentangling asr and mt errors in speech translation,” in MT Summit 2017, 2017. [OpenAIRE]
 N. Ruiz, M. A. Di Gangi, N. Bertoldi, and M. Federico, “Assessing the tolerance of neural machine translation systems against speech recognition errors.” in INTERSPEECH, 2017, pp. 2635-2639.