
We introduce a sleep-wake architecture for lifelong conversational memory in local language models running on consumer hardware. During wake, the system extracts facts from conversation and stores them in context. During sleep, it consolidates these facts into model weights via LoRA fine-tuning using spaced-repetition-inspired training data. We validate on a 3B parameter model (Llama-3.2-3B-Instruct-4bit) running on an 8GB MacBook Air M3, demonstrating that sleep cycles produce measurable memory formation with a narrow viable learning rate window (~1e-4) and a spaced repetition effect where repeated sleep cycles improve recall. This establishes the basic feasibility of sleep-wake memory consolidation in local LLMs.
Part of the Sleeping LLM research series on sleep-wake memory consolidation for lifelong learning in language models.
memory consolidation, language models, lifelong learning, LoRA, continual learning, sleep-wake cycle
memory consolidation, language models, lifelong learning, LoRA, continual learning, sleep-wake cycle
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
