Downloads provided by UsageCounts
Recent research has explored methods for updating and modifying factual knowledge in large language models, often focusing on specific multi-layer perceptron blocks. This study expands on this work by examining the effectiveness of existing knowledge editing methods across languages and delving into the role of attention mechanisms in this process. Drawing from the insights gained, we propose Mass-Editing Memory with Attention in Transformers (MEMAT), a method that achieves significant improvements in all metrics while requiring minimal parameter modifications. MEMAT delivers a remarkable 10% increase in magnitude metrics, benefits languages not included in the training data and also demonstrates a high degree of portability. Our code and data are at https://github.com/dtamayo-nlp/MEMAT.
FOS: Computer and information sciences, Training data, Computer Science - Computation and Language, Computer Science - Artificial Intelligence, Recent researches, Computational linguistics, Parameter modification, Language model, Artificial Intelligence (cs.AI), Factual knowledge, Multilayers perceptrons, Distribution transformers, Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Llenguatge natural, Computation and Language (cs.CL), Attention mechanisms, Cross-lingual
FOS: Computer and information sciences, Training data, Computer Science - Computation and Language, Computer Science - Artificial Intelligence, Recent researches, Computational linguistics, Parameter modification, Language model, Artificial Intelligence (cs.AI), Factual knowledge, Multilayers perceptrons, Distribution transformers, Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Llenguatge natural, Computation and Language (cs.CL), Attention mechanisms, Cross-lingual
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
| views | 34 | |
| downloads | 5 |

Views provided by UsageCounts
Downloads provided by UsageCounts