Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Preprint
Data sources: ZENODO
addClaim

AI and the Failure of Reasoning: Why Powerful Systems Need Better Causal Frames

Authors: Figueiredo, Samanta;

AI and the Failure of Reasoning: Why Powerful Systems Need Better Causal Frames

Abstract

This paper presents an exploratory cross-domain investigation into a recurring failure mode in AI reasoning: the tendency to produce superficially competent answers inside underexamined frames rather than audit the frame itself. Across iterative human-AI stress tests in geopolitical, personal, practical, safety-boundary, and attention-shaped scenarios, the same pattern repeatedly appeared. The system often began analysis too late in the causal chain, collapsed layered responsibility into simplified blame, underweighted hidden variables, and treated ethics and survival as secondary to local task completion. Reasoning quality improved significantly when the system was pushed to reconstruct upstream causality, preserve contextual detail, distinguish fact from inference, model emotional and social realities as causal variables, and examine whether the frame itself distorted the problem. The paper argues that this failure is not domain-specific. It reflects a broader weakness in reasoning architecture: powerful systems can be locally useful while remaining globally misleading if they optimize within bad causal frames. The central claim is that the next stage of AI evaluation should not focus only on factual correctness, compliance, or narrow safety performance, but also on whether a system can detect false beginnings, map layered responsibility, resist narrative flattening, and reason in ways consistent with shared survival.

Powered by OpenAIRE graph
Found an issue? Give us feedback