Abstract
Research reveals that large language models exhibit fragile reasoning capabilities when subjected to perturbations, with open-weight models showing significant accuracy drops and evidence of memory pollution in dense attention mechanisms.
While Large Language Models (LLMs) achieve high performance on standard mathematical benchmarks, their underlying reasoning processes remain highly overfit to standard textual formatting. We propose a perturbation pipeline consisting of 14 techniques to evaluate robustness of LLM reasoning. We apply this pipeline to AIME 2024 dataset and evalute 8 state-of-the-art models on the resulting benchmark. While frontier models exhibit resilience, open weights reasoning models suffer catastrophic collapses (up to 55% average accuracy drops across perturbations and up to 100% on some), exposing structural fragility. To further disentangle mechanical parsing failures from downstream reasoning failures, we strictly isolate the models' working memory capacity by forcing models to solve multiple unperturbed mathematical problems sequentially within a single context window. Our results indicate that open weight models ranging from 7B to 120B parameters and Claude Opus 4.6 exhibit accuracy decay on subsequent problems. This degradation demonstrates that intermediate reasoning steps permanently pollute standard dense attention mechanisms. We argue that to achieve reliable reasoning, future reasoning architectures must integrate explicit contextual resets within a model's own Chain-of-Thought, leading to fundamental open questions regarding the optimal granularity of atomic reasoning tasks.
Community
We introduce the Robust Reasoning Benchmark (RRB) using 14 meaning-preserving textual perturbations to expose LLM reasoning fragility. Additionally, we leverage sequential experiments to demonstrate that models suffer from attention degradation inside of a single chain of thought prompt. This raises a fundamental open question: what is the optimal granularity of a reasoning task before a model's dense attention starts degrading its reasoning inside of a single prompt?
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Beyond Content Safety: Real-Time Monitoring for Reasoning Vulnerabilities in Large Language Models (2026)
- Same Meaning, Different Scores: Lexical and Syntactic Sensitivity in LLM Evaluation (2026)
- Broken Chains: The Cost of Incomplete Reasoning in LLMs (2026)
- Robustness and Reasoning Fidelity of Large Language Models in Long-Context Code Question Answering (2026)
- X-RAY: Mapping LLM Reasoning Capability via Formalized and Calibrated Probes (2026)
- Semantic Invariance in Agentic AI (2026)
- LogicGraph : Benchmarking Multi-Path Logical Reasoning via Neuro-Symbolic Generation and Verification (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.08571 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper