Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Preprint . 2026
Data sources: ZENODO
addClaim

EROS-1: An Identity-Stability Kernel for Salience-Preserving and Risk-Proportionate LLM Interaction

Authors: Valente, Stefano;

EROS-1: An Identity-Stability Kernel for Salience-Preserving and Risk-Proportionate LLM Interaction

Abstract

Large Language Models (LLMs) deployed in sustained interactive settings exhibit progressive salience attenuation, over-accommodation, and loss of epistemic friction. These effects are driven less by model architecture than by interaction regimes that lack identity consolidation, verification, and proportional capability control. We introduce EROS-1, a progressive and auditable user kernel that stabilizes long-horizon interaction by incrementally consolidating verified user information and gating model capabilities as a function of identity stability. Kernel stability is formally defined using entropy-based measures and quantitative contradiction penalties, estimated via shadow interaction data to ensure non-reactive optimization.EROS-1 preserves exploratory interaction while enabling high-trust reasoning only under demonstrable stability, and is explicitly designed to align with the European Union AI Act’s principles of risk proportionality, auditability, and non-manipulative adaptation. Empirical evaluation using shadow trials, ANCOVA, and chi-square analysis demonstrates significant improvements in salience retention and capability stability over baseline interaction regimes.

Keywords

Large Language Models; User Modeling; Identity Stability; Salience Preservation; AI Governance; EU AI Act; Auditability; Risk-Proportionate AI; Long-Horizon Interaction

Powered by OpenAIRE graph
Found an issue? Give us feedback