
We introduce a dual-system memory architecture for language models inspired by Complementary Learning Systems (CLS) theory. MEMIT (Mass-Editing Memory in Transformers) serves as fast hippocampal encoding, injecting facts directly into MLP weights during wake. LoRA fine-tuning serves as slow neocortical consolidation during sleep. We develop covariance-regularized MEMIT with cross-edit null-space constraints that prevent new edits from overwriting previous ones, and validate the dual system across 3B, 8B, and 70B parameter models. Key ablations show: (1) the dual system outperforms either component alone, (2) null-space constraints achieve perfect retention across sequential edits, and (3) the Woodbury identity enables efficient covariance regularization in N x N space rather than d x d.
Part of the Sleeping LLM research series on sleep-wake memory consolidation for lifelong learning in language models.
language models, null-space constraints, dual-system memory, lifelong learning, complementary learning systems, knowledge editing, MEMIT
language models, null-space constraints, dual-system memory, lifelong learning, complementary learning systems, knowledge editing, MEMIT
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
