
Non-autoregressive models generate target words in a parallel way, which achieve a faster decoding speed but at the sacrifice of translation accuracy. To remedy a flawed translation by non-autoregressive models, a promising approach is to train a conditional masked translation model (CMTM), and refine the generated results within several iterations. Unfortunately, such approach hardly considers the \textit{sequential dependency} among target words, which inevitably results in a translation degradation. Hence, instead of solely training a Transformer-based CMTM, we propose a Self-Review Mechanism to infuse sequential information into it. Concretely, we insert a left-to-right mask to the same decoder of CMTM, and then induce it to autoregressively review whether each generated word from CMTM is supposed to be replaced or kept. The experimental results (WMT14 En$\leftrightarrow$De and WMT16 En$\leftrightarrow$Ro) demonstrate that our model uses dramatically less training computations than the typical CMTM, as well as outperforms several state-of-the-art non-autoregressive models by over 1 BLEU. Through knowledge distillation, our model even surpasses a typical left-to-right Transformer model, while significantly speeding up decoding.
accepted to coling 2020
Syntax-based Translation Models, FOS: Computer and information sciences, Neural Machine Translation, Artificial intelligence, Translation (biology), Speech recognition, Autoregressive model, Biochemistry, Quantum mechanics, Gene, Visual Question Answering in Images and Videos, FOS: Economics and business, Machine Translation, Artificial Intelligence, FOS: Mathematics, Econometrics, Machine translation, Natural Language Processing, Transformer, Computer Science - Computation and Language, Topic Modeling, Natural language processing, Physics, Messenger RNA, Voltage, Statistical Machine Translation and Natural Language Processing, Computer science, Language Modeling, Algorithm, Chemistry, Computer Science, Physical Sciences, Computation, Dependency (UML), Computer Vision and Pattern Recognition, Decoding methods, Computation and Language (cs.CL), Mathematics
Syntax-based Translation Models, FOS: Computer and information sciences, Neural Machine Translation, Artificial intelligence, Translation (biology), Speech recognition, Autoregressive model, Biochemistry, Quantum mechanics, Gene, Visual Question Answering in Images and Videos, FOS: Economics and business, Machine Translation, Artificial Intelligence, FOS: Mathematics, Econometrics, Machine translation, Natural Language Processing, Transformer, Computer Science - Computation and Language, Topic Modeling, Natural language processing, Physics, Messenger RNA, Voltage, Statistical Machine Translation and Natural Language Processing, Computer science, Language Modeling, Algorithm, Chemistry, Computer Science, Physical Sciences, Computation, Dependency (UML), Computer Vision and Pattern Recognition, Decoding methods, Computation and Language (cs.CL), Mathematics
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 3 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
