
Injecting external knowledge can improve the performance of pre-trained language models (PLMs) on various downstream NLP tasks. However, massive retraining is required to deploy new knowledge injection methods or knowledge bases for downstream tasks. In this work, we are the first to study how to improve the flexibility and efficiency of knowledge injection by reusing existing downstream models. To this end, we explore a new paradigm plug-and-play knowledge injection, where knowledge bases are injected into frozen existing downstream models by a knowledge plugin. Correspondingly, we propose a plug-and-play injection method map-tuning, which trains a mapping of knowledge embeddings to enrich model inputs with mapped embeddings while keeping model parameters frozen. Experimental results on three knowledge-driven NLP tasks show that existing injection methods are not suitable for the new paradigm, while map-tuning effectively improves the performance of downstream models. Moreover, we show that a frozen downstream model can be well adapted to different domains with different mapping networks of domain knowledge. Our code and models are available at https://github.com/THUNLP/Knowledge-Plugin.
ACL 2023
Syntax-based Translation Models, FOS: Computer and information sciences, Artificial intelligence, History, China, Computer Science - Computation and Language, Volume (thermodynamics), Natural language processing, Physics, Pretrained Models, Computational linguistics, Statistical Machine Translation and Natural Language Processing, Computer science, Quantum mechanics, Language Modeling, Machine Translation, Archaeology, Artificial Intelligence, Part-of-Speech Tagging, Computer Science, Physical Sciences, Zhàng, Computation and Language (cs.CL), Natural Language Processing
Syntax-based Translation Models, FOS: Computer and information sciences, Artificial intelligence, History, China, Computer Science - Computation and Language, Volume (thermodynamics), Natural language processing, Physics, Pretrained Models, Computational linguistics, Statistical Machine Translation and Natural Language Processing, Computer science, Quantum mechanics, Language Modeling, Machine Translation, Archaeology, Artificial Intelligence, Part-of-Speech Tagging, Computer Science, Physical Sciences, Zhàng, Computation and Language (cs.CL), Natural Language Processing
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 6 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
