publication . Preprint . Other literature type . Conference object . 2019

Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training

Cheng, Qiao; Fan, Meiyuan; Han, Yaqian; Huang, Jin; Duan, Yitao;
Open Access English
  • Published: 02 Nov 2019
Abstract
Comment: Accepted at the 16th International Workshop on Spoken Language Translation (IWSLT 2019)
Subjects
free text keywords: Computer Science - Computation and Language, Electrical Engineering and Systems Science - Audio and Speech Processing
Download fromView all 4 versions
Zenodo
Other literature type . 2019
Provider: Datacite
Zenodo
Other literature type . 2019
Provider: Datacite
ZENODO
Conference object . 2019
Provider: ZENODO
27 references, page 1 of 2

[1] R. J. Weiss, J. Chorowski, N. Jaitly, Y. Wu, and Z. Chen, “Sequence-to-sequence models can directly translate foreign speech,” Proc. Interspeech 2017, pp. 2625-2629, 2017. [OpenAIRE]

[2] A. Bérard, O. Pietquin, L. Besacier, and C. Servan, “Listen and translate: A proof of concept for end-toend speech-to-text translation,” in NIPS Workshop on end-to-end learning for speech and audio processing, 2016. [OpenAIRE]

[3] S. Peitz, S. Wiesler, M. Nußbaum-Thom, and H. Ney, “Spoken language translation using automatically transcribed text in training,” in International Workshop on Spoken Language Translation (IWSLT) 2012, 2012. [OpenAIRE]

[4] M. Sperber, G. Neubig, J. Niehues, and A. Waibel, “Neural lattice-to-sequence models for uncertain inputs,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017, pp. 1380-1389. [OpenAIRE]

[5] M. Post, G. Kumar, A. Lopez, D. Karakos, C. CallisonBurch, and S. Khudanpur, “Improved speech-to-text translation with the fisher and callhome spanishenglish speech translation corpus,” in Proc. IWSLT, 2013.

[6] N. Jan, R. Cattoni, S. Sebastian, M. Cettolo, M. Turchi, and M. Federico, “The iwslt 2018 evaluation campaign,” in International Workshop on Spoken Language Translation, 2018, pp. 2-6.

[7] X. Li, H. Xue, W. Chen, Y. Liu, Y. Feng, and Q. Liu, “Improving the robustness of speech translation,” arXiv preprint arXiv:1811.00728, 2018.

[8] M. Sperber, J. Niehues, and A. Waibel, “Toward robust neural machine translation for noisy input sequences,” in International Workshop on Spoken Language Translation (IWSLT), Tokyo, Japan, 2017. [OpenAIRE]

[9] A. Bérard, L. Besacier, A. C. Kocabiyikoglu, and O. Pietquin, “End-to-end automatic speech translation of audiobooks,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 6224-6228. [OpenAIRE]

[10] D. Serdyuk, Y. Wang, C. Fuegen, A. Kumar, B. Liu, and Y. Bengio, “Towards end-to-end spoken language understanding,” in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 5754-5758. [OpenAIRE]

[11] L. Duong, A. Anastasopoulos, D. Chiang, S. Bird, and T. Cohn, “An attentional model for speech translation without transcription,” in Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2016, pp. 949-959. [OpenAIRE]

[12] S. Bansal, H. Kamper, A. Lopez, and S. Goldwater, “Towards speech-to-text translation without speech recognition,” EACL 2017, p. 474, 2017.

[13] Y. Belinkov and Y. Bisk, “Synthetic and natural noise both break neural machine translation,” arXiv preprint arXiv:1711.02173, 2017. [OpenAIRE]

[14] N.-T. Le, B. Lecouteux, and L. Besacier, “Disentangling asr and mt errors in speech translation,” in MT Summit 2017, 2017. [OpenAIRE]

[15] N. Ruiz, M. A. Di Gangi, N. Bertoldi, and M. Federico, “Assessing the tolerance of neural machine translation systems against speech recognition errors.” in INTERSPEECH, 2017, pp. 2635-2639.

27 references, page 1 of 2
Powered by OpenAIRE Research Graph
Any information missing or wrong?Report an Issue