
Continuous Representations, Discrete Commitment: A Causal Threshold in Decoder-Only LLMs Correlational and interventional analyses of LLM internals appear to disagree: probes show gradual representational change across depth, while activation patching reveals sharp behavioral transitions. We resolve this by showing the two methods measure different properties. We perform layerwise residual-stream swaps with paired controls across three decoder-only architectures (GPT-2 Small, Gemma-2-2B, Qwen2.5-1.5B) and find a replicated causal commitment transition at 62–71% network depth. Below this threshold, swaps produce negligible behavioral change; at or above it, outputs flip immediately with large margin transfer. The transition is specific to the main intervention (not matched by random-norm, self, or position-shuffle controls) and stable across patch scales and random seeds in the two mid-size models. Representations evolve continuously. Causal commitment does not. The two findings are compatible once the distinction between representational change and output determination is made explicit. Code and evaluation notebooks are available in the companion repository. Keywords: mechanistic interpretability, activation patching, causal intervention, commitment threshold, decoder-only transformers
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
